var/home/core/zuul-output/0000755000175000017500000000000015111275235014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111302556015470 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004647052515111302547017711 0ustar rootrootNov 25 09:48:21 crc systemd[1]: Starting Kubernetes Kubelet... Nov 25 09:48:21 crc restorecon[4754]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:21 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:22 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 25 09:48:23 crc restorecon[4754]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 25 09:48:23 crc kubenswrapper[4926]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:48:23 crc kubenswrapper[4926]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 25 09:48:23 crc kubenswrapper[4926]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:48:23 crc kubenswrapper[4926]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:48:23 crc kubenswrapper[4926]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 25 09:48:23 crc kubenswrapper[4926]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.760300 4926 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765806 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765824 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765828 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765833 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765840 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765844 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765849 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765853 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765858 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765862 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765865 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765869 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765873 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765878 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765882 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765886 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765889 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765893 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765897 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765901 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765905 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765909 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765914 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765919 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765923 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765944 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765949 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765953 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765957 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765961 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765965 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765970 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765974 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765978 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765982 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765986 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765989 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.765994 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766000 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766005 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766011 4926 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766016 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766020 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766025 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766030 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766034 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766038 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766042 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766046 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766050 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766054 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766058 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766062 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766065 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766069 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766073 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766076 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766080 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766085 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766089 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766094 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766098 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766102 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766106 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766109 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766113 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766117 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766122 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766127 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766131 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.766137 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766232 4926 flags.go:64] FLAG: --address="0.0.0.0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766242 4926 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766252 4926 flags.go:64] FLAG: --anonymous-auth="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766259 4926 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766265 4926 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766269 4926 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766275 4926 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766281 4926 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766285 4926 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766290 4926 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766295 4926 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766336 4926 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766342 4926 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766347 4926 flags.go:64] FLAG: --cgroup-root="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766351 4926 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766355 4926 flags.go:64] FLAG: --client-ca-file="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766360 4926 flags.go:64] FLAG: --cloud-config="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766364 4926 flags.go:64] FLAG: --cloud-provider="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766368 4926 flags.go:64] FLAG: --cluster-dns="[]" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766374 4926 flags.go:64] FLAG: --cluster-domain="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766378 4926 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766383 4926 flags.go:64] FLAG: --config-dir="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766388 4926 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766392 4926 flags.go:64] FLAG: --container-log-max-files="5" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766399 4926 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766403 4926 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766408 4926 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766412 4926 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766417 4926 flags.go:64] FLAG: --contention-profiling="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766422 4926 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766426 4926 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766432 4926 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766436 4926 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766442 4926 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766447 4926 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766451 4926 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766456 4926 flags.go:64] FLAG: --enable-load-reader="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766460 4926 flags.go:64] FLAG: --enable-server="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766464 4926 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766471 4926 flags.go:64] FLAG: --event-burst="100" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766475 4926 flags.go:64] FLAG: --event-qps="50" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766480 4926 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766484 4926 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766489 4926 flags.go:64] FLAG: --eviction-hard="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766494 4926 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766499 4926 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766503 4926 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766507 4926 flags.go:64] FLAG: --eviction-soft="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766512 4926 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766516 4926 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766520 4926 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766525 4926 flags.go:64] FLAG: --experimental-mounter-path="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766529 4926 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766533 4926 flags.go:64] FLAG: --fail-swap-on="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766538 4926 flags.go:64] FLAG: --feature-gates="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766543 4926 flags.go:64] FLAG: --file-check-frequency="20s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766548 4926 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766553 4926 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766557 4926 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766562 4926 flags.go:64] FLAG: --healthz-port="10248" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766567 4926 flags.go:64] FLAG: --help="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766572 4926 flags.go:64] FLAG: --hostname-override="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766576 4926 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766580 4926 flags.go:64] FLAG: --http-check-frequency="20s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766585 4926 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766589 4926 flags.go:64] FLAG: --image-credential-provider-config="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766593 4926 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766598 4926 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766602 4926 flags.go:64] FLAG: --image-service-endpoint="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766606 4926 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766610 4926 flags.go:64] FLAG: --kube-api-burst="100" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766615 4926 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766619 4926 flags.go:64] FLAG: --kube-api-qps="50" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766623 4926 flags.go:64] FLAG: --kube-reserved="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766628 4926 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766631 4926 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766636 4926 flags.go:64] FLAG: --kubelet-cgroups="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766640 4926 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766645 4926 flags.go:64] FLAG: --lock-file="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766650 4926 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766654 4926 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766658 4926 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766665 4926 flags.go:64] FLAG: --log-json-split-stream="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766670 4926 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766674 4926 flags.go:64] FLAG: --log-text-split-stream="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766678 4926 flags.go:64] FLAG: --logging-format="text" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766683 4926 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766688 4926 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766692 4926 flags.go:64] FLAG: --manifest-url="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766696 4926 flags.go:64] FLAG: --manifest-url-header="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766702 4926 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766706 4926 flags.go:64] FLAG: --max-open-files="1000000" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766712 4926 flags.go:64] FLAG: --max-pods="110" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766716 4926 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766721 4926 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766725 4926 flags.go:64] FLAG: --memory-manager-policy="None" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766729 4926 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766737 4926 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766741 4926 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766746 4926 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766756 4926 flags.go:64] FLAG: --node-status-max-images="50" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766760 4926 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766765 4926 flags.go:64] FLAG: --oom-score-adj="-999" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766769 4926 flags.go:64] FLAG: --pod-cidr="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766773 4926 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766781 4926 flags.go:64] FLAG: --pod-manifest-path="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766786 4926 flags.go:64] FLAG: --pod-max-pids="-1" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766790 4926 flags.go:64] FLAG: --pods-per-core="0" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766795 4926 flags.go:64] FLAG: --port="10250" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766801 4926 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766806 4926 flags.go:64] FLAG: --provider-id="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766811 4926 flags.go:64] FLAG: --qos-reserved="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766816 4926 flags.go:64] FLAG: --read-only-port="10255" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766821 4926 flags.go:64] FLAG: --register-node="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766826 4926 flags.go:64] FLAG: --register-schedulable="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766832 4926 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766842 4926 flags.go:64] FLAG: --registry-burst="10" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766848 4926 flags.go:64] FLAG: --registry-qps="5" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766853 4926 flags.go:64] FLAG: --reserved-cpus="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766858 4926 flags.go:64] FLAG: --reserved-memory="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766865 4926 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766870 4926 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766875 4926 flags.go:64] FLAG: --rotate-certificates="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766880 4926 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766885 4926 flags.go:64] FLAG: --runonce="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766890 4926 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766896 4926 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766901 4926 flags.go:64] FLAG: --seccomp-default="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766906 4926 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766911 4926 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766918 4926 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766923 4926 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766928 4926 flags.go:64] FLAG: --storage-driver-password="root" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766959 4926 flags.go:64] FLAG: --storage-driver-secure="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766964 4926 flags.go:64] FLAG: --storage-driver-table="stats" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766968 4926 flags.go:64] FLAG: --storage-driver-user="root" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766972 4926 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766977 4926 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766981 4926 flags.go:64] FLAG: --system-cgroups="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766986 4926 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766993 4926 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.766997 4926 flags.go:64] FLAG: --tls-cert-file="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767001 4926 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767007 4926 flags.go:64] FLAG: --tls-min-version="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767011 4926 flags.go:64] FLAG: --tls-private-key-file="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767015 4926 flags.go:64] FLAG: --topology-manager-policy="none" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767019 4926 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767023 4926 flags.go:64] FLAG: --topology-manager-scope="container" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767027 4926 flags.go:64] FLAG: --v="2" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767034 4926 flags.go:64] FLAG: --version="false" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767039 4926 flags.go:64] FLAG: --vmodule="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767046 4926 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767050 4926 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767163 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767168 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767173 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767178 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767183 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767187 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767191 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767195 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767199 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767205 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767209 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767212 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767216 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767220 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767223 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767227 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767232 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767237 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767242 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767247 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767252 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767256 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767259 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767263 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767268 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767271 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767275 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767279 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767283 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767287 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767290 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767294 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767297 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767301 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767306 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767309 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767313 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767317 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767320 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767324 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767327 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767333 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767336 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767340 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767344 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767347 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767351 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767355 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767358 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767363 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767367 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767371 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767374 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767378 4926 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767382 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767386 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767390 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767393 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767398 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767401 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767405 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767409 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767412 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767417 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767422 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767427 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767431 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767435 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767439 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767443 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.767448 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.767462 4926 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.775857 4926 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.775884 4926 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775964 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775971 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775976 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775980 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775984 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775988 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775991 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775995 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.775999 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776002 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776006 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776009 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776013 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776017 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776020 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776061 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776067 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776072 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776076 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776080 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776085 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776090 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776094 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776099 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776102 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776106 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776110 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776114 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776117 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776121 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776124 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776128 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776131 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776135 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776140 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776143 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776147 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776150 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776154 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776158 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776161 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776165 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776168 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776172 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776176 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776179 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776183 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776187 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776190 4926 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776195 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776199 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776202 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776206 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776209 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776213 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776226 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776230 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776233 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776237 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776241 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776244 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776249 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776254 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776258 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776262 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776266 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776270 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776274 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776277 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776281 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776285 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.776292 4926 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776412 4926 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776419 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776424 4926 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776428 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776432 4926 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776435 4926 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776439 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776443 4926 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776447 4926 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776451 4926 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776455 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776459 4926 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776462 4926 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776466 4926 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776469 4926 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776473 4926 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776477 4926 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776481 4926 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776484 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776488 4926 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776491 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776495 4926 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776499 4926 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776502 4926 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776506 4926 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776509 4926 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776513 4926 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776517 4926 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776521 4926 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776524 4926 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776528 4926 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776532 4926 feature_gate.go:330] unrecognized feature gate: Example Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776535 4926 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776539 4926 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776543 4926 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776546 4926 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776550 4926 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776554 4926 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776558 4926 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776562 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776567 4926 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776572 4926 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776576 4926 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776581 4926 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776586 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776590 4926 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776594 4926 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776598 4926 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776602 4926 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776605 4926 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776609 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776613 4926 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776617 4926 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776622 4926 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776626 4926 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776630 4926 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776633 4926 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776637 4926 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776641 4926 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776645 4926 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776649 4926 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776652 4926 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776656 4926 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776660 4926 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776665 4926 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776669 4926 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776673 4926 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776676 4926 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776680 4926 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776684 4926 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.776688 4926 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.776693 4926 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.780305 4926 server.go:940] "Client rotation is on, will bootstrap in background" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.800715 4926 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.800829 4926 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.802320 4926 server.go:997] "Starting client certificate rotation" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.802349 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.807211 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-07 08:26:56.592083027 +0000 UTC Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.807285 4926 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 286h38m32.784800337s for next certificate rotation Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.841716 4926 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.843539 4926 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.861454 4926 log.go:25] "Validated CRI v1 runtime API" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.900149 4926 log.go:25] "Validated CRI v1 image API" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.902355 4926 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.909836 4926 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-25-09-43-58-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.909886 4926 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.927278 4926 manager.go:217] Machine: {Timestamp:2025-11-25 09:48:23.92436043 +0000 UTC m=+0.608316877 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:612af39e-7a36-4010-be71-d81687d215a9 BootID:ae35d4ee-cf02-4f2c-87da-f07de9ee360a Filesystems:[{Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e8:44:90 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e8:44:90 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:9f:8d:64 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:bb:79:3e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:33:8c:d0 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:3f:24:74 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:c1:61:15 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:82:6b:ec:5b:61:74 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:76:9f:42:16:ed:43 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.927577 4926 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.927779 4926 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.928197 4926 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.928408 4926 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.928471 4926 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.928763 4926 topology_manager.go:138] "Creating topology manager with none policy" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.928778 4926 container_manager_linux.go:303] "Creating device plugin manager" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.929366 4926 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.929406 4926 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.930186 4926 state_mem.go:36] "Initialized new in-memory state store" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.930328 4926 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.933741 4926 kubelet.go:418] "Attempting to sync node with API server" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.933772 4926 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.933824 4926 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.933841 4926 kubelet.go:324] "Adding apiserver pod source" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.933857 4926 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.938451 4926 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.939557 4926 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.941993 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.942094 4926 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.942055 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:23 crc kubenswrapper[4926]: E1125 09:48:23.942181 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:23 crc kubenswrapper[4926]: E1125 09:48:23.942114 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943632 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943658 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943666 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943673 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943686 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943694 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943702 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943715 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943726 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943735 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943748 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.943755 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.944599 4926 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.945152 4926 server.go:1280] "Started kubelet" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.945253 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.945496 4926 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 25 09:48:23 crc systemd[1]: Started Kubernetes Kubelet. Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.949877 4926 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.950646 4926 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.952983 4926 server.go:460] "Adding debug handlers to kubelet server" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954340 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954397 4926 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954614 4926 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954632 4926 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954707 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 01:53:07.571768736 +0000 UTC Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954784 4926 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 472h4m43.616990559s for next certificate rotation Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.954825 4926 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 25 09:48:23 crc kubenswrapper[4926]: E1125 09:48:23.955002 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.955810 4926 factory.go:55] Registering systemd factory Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.955838 4926 factory.go:221] Registration of the systemd container factory successfully Nov 25 09:48:23 crc kubenswrapper[4926]: E1125 09:48:23.955109 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.90:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b36f4c4aa7efd default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 09:48:23.945109245 +0000 UTC m=+0.629065682,LastTimestamp:2025-11-25 09:48:23.945109245 +0000 UTC m=+0.629065682,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 09:48:23 crc kubenswrapper[4926]: E1125 09:48:23.961631 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="200ms" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.962084 4926 factory.go:153] Registering CRI-O factory Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.962130 4926 factory.go:221] Registration of the crio container factory successfully Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.962291 4926 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.962330 4926 factory.go:103] Registering Raw factory Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.962359 4926 manager.go:1196] Started watching for new ooms in manager Nov 25 09:48:23 crc kubenswrapper[4926]: W1125 09:48:23.962430 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:23 crc kubenswrapper[4926]: E1125 09:48:23.962522 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.963090 4926 manager.go:319] Starting recovery of all containers Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969751 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969824 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969843 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969859 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969873 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969886 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969902 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969916 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969933 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969969 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.969985 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970002 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970018 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970060 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970074 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970089 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970105 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970119 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970136 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970150 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970163 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970177 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970193 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970207 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970221 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970237 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970256 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970272 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970285 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970325 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970340 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970354 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970389 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970409 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970422 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970439 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970453 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970468 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970485 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970499 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970515 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970533 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970552 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970567 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970582 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970600 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970614 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970628 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970646 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970662 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970678 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970695 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970716 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970734 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970751 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970768 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970786 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970801 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970816 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970831 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970849 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970864 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970880 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970897 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970913 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.970925 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971014 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971030 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971041 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971054 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971067 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971079 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971091 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971102 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971113 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971127 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971138 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971150 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971161 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971173 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971210 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971222 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971238 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971250 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971261 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971273 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971285 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971297 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971313 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971328 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971344 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971357 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971369 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971381 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971394 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971408 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971422 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971436 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971449 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971461 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971473 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971486 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971497 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971512 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971528 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971541 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971554 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971569 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971581 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971595 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971609 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971622 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971637 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971651 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971663 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971673 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971685 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971697 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971710 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971722 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971733 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971743 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971754 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971766 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971778 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971791 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971813 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971825 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971839 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971851 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971864 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.971878 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.974229 4926 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.974292 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.974324 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.974345 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.974368 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.974391 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975330 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975359 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975388 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975404 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975417 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975428 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975442 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975461 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975475 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975489 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975503 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975514 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975526 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975537 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975549 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975565 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975574 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975586 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975596 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975606 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975619 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975630 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975640 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975652 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975663 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975675 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975685 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975695 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975704 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975717 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975728 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975739 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975751 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975763 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975775 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.975785 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976304 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976334 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976359 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976374 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976389 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976408 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976420 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976494 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976506 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976517 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976530 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976543 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976557 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976568 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976580 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976591 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976603 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976614 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976626 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976638 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976650 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976661 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.976671 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977225 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977253 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977269 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977285 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977301 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977314 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977325 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977340 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977351 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977364 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977379 4926 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977394 4926 reconstruct.go:97] "Volume reconstruction finished" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.977405 4926 reconciler.go:26] "Reconciler: start to sync state" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.981069 4926 manager.go:324] Recovery completed Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.994090 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.996107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.996164 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.996215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.997312 4926 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.997335 4926 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 25 09:48:23 crc kubenswrapper[4926]: I1125 09:48:23.997362 4926 state_mem.go:36] "Initialized new in-memory state store" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.007814 4926 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.010985 4926 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.011067 4926 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.011106 4926 kubelet.go:2335] "Starting kubelet main sync loop" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.011294 4926 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 25 09:48:24 crc kubenswrapper[4926]: W1125 09:48:24.015406 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.015496 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.016225 4926 policy_none.go:49] "None policy: Start" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.017079 4926 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.017109 4926 state_mem.go:35] "Initializing new in-memory state store" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.055124 4926 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.072591 4926 manager.go:334] "Starting Device Plugin manager" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.072670 4926 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.072687 4926 server.go:79] "Starting device plugin registration server" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.073369 4926 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.073404 4926 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.073579 4926 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.073748 4926 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.073763 4926 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.083483 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.111789 4926 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.111977 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.113271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.113314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.113326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.113516 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.113788 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.113833 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114211 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114219 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114332 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114499 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114553 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.114994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115079 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115176 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115209 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115799 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.115990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116246 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116294 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116428 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116643 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.116686 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.117646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.117671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.117682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.117831 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.117848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.117858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.162379 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="400ms" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.174359 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.175683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.175739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.175752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.175807 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.176626 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.90:6443: connect: connection refused" node="crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.180925 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.180974 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.180998 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181018 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181037 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181053 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181072 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181136 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181185 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181227 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181259 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181282 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181308 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.181372 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283226 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283306 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283330 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283347 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283379 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283423 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283492 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283514 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283541 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283560 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283546 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283628 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283584 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283813 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283886 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283925 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283981 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.283988 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284016 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284074 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284097 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284105 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284113 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284171 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284192 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284224 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.284171 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.377700 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.379127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.379171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.379182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.379207 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.379663 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.90:6443: connect: connection refused" node="crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.461917 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.478624 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.495785 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.516398 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: W1125 09:48:24.518059 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-9cb8e3c62d88664a7330fd32d463e7b3358e2dbdf2b38f22ae608a1d05b3bbaa WatchSource:0}: Error finding container 9cb8e3c62d88664a7330fd32d463e7b3358e2dbdf2b38f22ae608a1d05b3bbaa: Status 404 returned error can't find the container with id 9cb8e3c62d88664a7330fd32d463e7b3358e2dbdf2b38f22ae608a1d05b3bbaa Nov 25 09:48:24 crc kubenswrapper[4926]: W1125 09:48:24.520373 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-3e9d176a1f80abdea3713be991790f5580e6f415b06eb63298cb6f4a0a537d8c WatchSource:0}: Error finding container 3e9d176a1f80abdea3713be991790f5580e6f415b06eb63298cb6f4a0a537d8c: Status 404 returned error can't find the container with id 3e9d176a1f80abdea3713be991790f5580e6f415b06eb63298cb6f4a0a537d8c Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.523018 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:24 crc kubenswrapper[4926]: W1125 09:48:24.528017 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a28b7b52c7d48d82e8a6c6d8c0c8bcbd8417acfce456ccd43c323523039f1452 WatchSource:0}: Error finding container a28b7b52c7d48d82e8a6c6d8c0c8bcbd8417acfce456ccd43c323523039f1452: Status 404 returned error can't find the container with id a28b7b52c7d48d82e8a6c6d8c0c8bcbd8417acfce456ccd43c323523039f1452 Nov 25 09:48:24 crc kubenswrapper[4926]: W1125 09:48:24.534122 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-1e75b541451b2a51917cb9737f1823d0a2e02bdf1d086ec8cb59eca13fc5ce47 WatchSource:0}: Error finding container 1e75b541451b2a51917cb9737f1823d0a2e02bdf1d086ec8cb59eca13fc5ce47: Status 404 returned error can't find the container with id 1e75b541451b2a51917cb9737f1823d0a2e02bdf1d086ec8cb59eca13fc5ce47 Nov 25 09:48:24 crc kubenswrapper[4926]: W1125 09:48:24.536122 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-d3dea86ba0964236e7dcff377f795c713fd973d71e5c6b7780572cc4f82d354b WatchSource:0}: Error finding container d3dea86ba0964236e7dcff377f795c713fd973d71e5c6b7780572cc4f82d354b: Status 404 returned error can't find the container with id d3dea86ba0964236e7dcff377f795c713fd973d71e5c6b7780572cc4f82d354b Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.564076 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="800ms" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.780235 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.781983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.782043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.782057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.782092 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:24 crc kubenswrapper[4926]: E1125 09:48:24.782585 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.90:6443: connect: connection refused" node="crc" Nov 25 09:48:24 crc kubenswrapper[4926]: I1125 09:48:24.946358 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.016490 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d3dea86ba0964236e7dcff377f795c713fd973d71e5c6b7780572cc4f82d354b"} Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.018012 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"1e75b541451b2a51917cb9737f1823d0a2e02bdf1d086ec8cb59eca13fc5ce47"} Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.018930 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a28b7b52c7d48d82e8a6c6d8c0c8bcbd8417acfce456ccd43c323523039f1452"} Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.020518 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3e9d176a1f80abdea3713be991790f5580e6f415b06eb63298cb6f4a0a537d8c"} Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.021600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9cb8e3c62d88664a7330fd32d463e7b3358e2dbdf2b38f22ae608a1d05b3bbaa"} Nov 25 09:48:25 crc kubenswrapper[4926]: W1125 09:48:25.179375 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:25 crc kubenswrapper[4926]: E1125 09:48:25.179829 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:25 crc kubenswrapper[4926]: W1125 09:48:25.196068 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:25 crc kubenswrapper[4926]: E1125 09:48:25.196211 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:25 crc kubenswrapper[4926]: W1125 09:48:25.221077 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:25 crc kubenswrapper[4926]: E1125 09:48:25.221190 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:25 crc kubenswrapper[4926]: E1125 09:48:25.365314 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="1.6s" Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.582831 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.584595 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.584639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.584655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.584691 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:25 crc kubenswrapper[4926]: E1125 09:48:25.585268 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.90:6443: connect: connection refused" node="crc" Nov 25 09:48:25 crc kubenswrapper[4926]: W1125 09:48:25.616631 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:25 crc kubenswrapper[4926]: E1125 09:48:25.616735 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:25 crc kubenswrapper[4926]: I1125 09:48:25.946692 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.026172 4926 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336" exitCode=0 Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.026289 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.026331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.027237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.027283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.027294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.028664 4926 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf" exitCode=0 Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.028733 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.028754 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.029441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.029500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.029513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.031893 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.031927 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.031956 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.031971 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.031982 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.032981 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.033004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.033014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.034563 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8" exitCode=0 Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.034665 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.034706 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.035363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.035391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.035402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.036545 4926 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828" exitCode=0 Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.036620 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.036635 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828"} Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.037239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.037279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.037289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.038589 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.039659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.039697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.039709 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:26 crc kubenswrapper[4926]: I1125 09:48:26.946718 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:26 crc kubenswrapper[4926]: E1125 09:48:26.967007 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="3.2s" Nov 25 09:48:27 crc kubenswrapper[4926]: W1125 09:48:27.007790 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:27 crc kubenswrapper[4926]: E1125 09:48:27.007905 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.046318 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.046783 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"a386f401a9684c2ea8c4941f2f3bbb17051796ccfe35e00841bb3426bc531e3f"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.047373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.047401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.047411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.051500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.051550 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.051566 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.051584 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.055153 4926 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14" exitCode=0 Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.055213 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.055328 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.056623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.056688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.056702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.063030 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.063727 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.063753 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.063828 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.063849 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f"} Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.064599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.064635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.064647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.064678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.064711 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.064721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.185790 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.187649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.187686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.187694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.187723 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:27 crc kubenswrapper[4926]: E1125 09:48:27.188700 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.90:6443: connect: connection refused" node="crc" Nov 25 09:48:27 crc kubenswrapper[4926]: W1125 09:48:27.369905 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.90:6443: connect: connection refused Nov 25 09:48:27 crc kubenswrapper[4926]: E1125 09:48:27.370007 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.90:6443: connect: connection refused" logger="UnhandledError" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.857308 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.863696 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:27 crc kubenswrapper[4926]: I1125 09:48:27.977310 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.069051 4926 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da" exitCode=0 Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.069128 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da"} Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.069823 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.071983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.072186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.072357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.073251 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be"} Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.073382 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.073458 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.073894 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.074881 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.074364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.074997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.075014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.074379 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.075056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.075069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.073386 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.076302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.076359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.076369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.076316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.076487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.076501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:28 crc kubenswrapper[4926]: I1125 09:48:28.665165 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.014586 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.079953 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00"} Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080030 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3"} Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080047 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa"} Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080062 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb"} Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080113 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080119 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080215 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080114 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.080360 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081176 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081889 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.081978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:29 crc kubenswrapper[4926]: I1125 09:48:29.992505 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.086701 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.086757 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.086770 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.086743 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b"} Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.087718 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.087747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.087757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.087719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.087848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.087860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.388956 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.390364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.390417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.390427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:30 crc kubenswrapper[4926]: I1125 09:48:30.390460 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.089133 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.089192 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.089265 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.090374 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.090422 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.090432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.090541 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.090573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.090591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:31 crc kubenswrapper[4926]: I1125 09:48:31.813518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.091596 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.092855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.092919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.092959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.098542 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.098686 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.099565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.099697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:32 crc kubenswrapper[4926]: I1125 09:48:32.099713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:34 crc kubenswrapper[4926]: E1125 09:48:34.083656 4926 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 25 09:48:34 crc kubenswrapper[4926]: I1125 09:48:34.120831 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:34 crc kubenswrapper[4926]: I1125 09:48:34.121100 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:34 crc kubenswrapper[4926]: I1125 09:48:34.122647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:34 crc kubenswrapper[4926]: I1125 09:48:34.122694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:34 crc kubenswrapper[4926]: I1125 09:48:34.122710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:34 crc kubenswrapper[4926]: I1125 09:48:34.984113 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:35 crc kubenswrapper[4926]: I1125 09:48:35.099429 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:35 crc kubenswrapper[4926]: I1125 09:48:35.101498 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:35 crc kubenswrapper[4926]: I1125 09:48:35.101552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:35 crc kubenswrapper[4926]: I1125 09:48:35.101562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:35 crc kubenswrapper[4926]: I1125 09:48:35.103571 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:36 crc kubenswrapper[4926]: I1125 09:48:36.102267 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:36 crc kubenswrapper[4926]: I1125 09:48:36.103448 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:36 crc kubenswrapper[4926]: I1125 09:48:36.103494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:36 crc kubenswrapper[4926]: I1125 09:48:36.103507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.754537 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.754782 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.756495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.756548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.756558 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.947251 4926 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.985016 4926 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 25 09:48:37 crc kubenswrapper[4926]: I1125 09:48:37.985303 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Nov 25 09:48:38 crc kubenswrapper[4926]: W1125 09:48:38.181597 4926 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 25 09:48:38 crc kubenswrapper[4926]: I1125 09:48:38.181752 4926 trace.go:236] Trace[938958197]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:48:28.180) (total time: 10001ms): Nov 25 09:48:38 crc kubenswrapper[4926]: Trace[938958197]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (09:48:38.181) Nov 25 09:48:38 crc kubenswrapper[4926]: Trace[938958197]: [10.001055753s] [10.001055753s] END Nov 25 09:48:38 crc kubenswrapper[4926]: E1125 09:48:38.181789 4926 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 25 09:48:38 crc kubenswrapper[4926]: I1125 09:48:38.360033 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 09:48:38 crc kubenswrapper[4926]: I1125 09:48:38.360137 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 09:48:38 crc kubenswrapper[4926]: I1125 09:48:38.367719 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 25 09:48:38 crc kubenswrapper[4926]: I1125 09:48:38.367813 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 25 09:48:39 crc kubenswrapper[4926]: I1125 09:48:39.021712 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]log ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]etcd ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/generic-apiserver-start-informers ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/priority-and-fairness-filter ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-apiextensions-informers ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-apiextensions-controllers ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/crd-informer-synced ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-system-namespaces-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 25 09:48:39 crc kubenswrapper[4926]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 25 09:48:39 crc kubenswrapper[4926]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/bootstrap-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/start-kube-aggregator-informers ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-registration-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-discovery-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]autoregister-completion ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-openapi-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 25 09:48:39 crc kubenswrapper[4926]: livez check failed Nov 25 09:48:39 crc kubenswrapper[4926]: I1125 09:48:39.021803 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:48:43 crc kubenswrapper[4926]: E1125 09:48:43.351334 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.353797 4926 trace.go:236] Trace[1679483970]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:48:28.484) (total time: 14869ms): Nov 25 09:48:43 crc kubenswrapper[4926]: Trace[1679483970]: ---"Objects listed" error: 14869ms (09:48:43.353) Nov 25 09:48:43 crc kubenswrapper[4926]: Trace[1679483970]: [14.869276493s] [14.869276493s] END Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.353843 4926 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 09:48:43 crc kubenswrapper[4926]: E1125 09:48:43.354914 4926 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.356540 4926 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.356579 4926 trace.go:236] Trace[56398666]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:48:32.554) (total time: 10802ms): Nov 25 09:48:43 crc kubenswrapper[4926]: Trace[56398666]: ---"Objects listed" error: 10802ms (09:48:43.356) Nov 25 09:48:43 crc kubenswrapper[4926]: Trace[56398666]: [10.802209478s] [10.802209478s] END Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.356610 4926 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.356874 4926 trace.go:236] Trace[908235131]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (25-Nov-2025 09:48:31.207) (total time: 12149ms): Nov 25 09:48:43 crc kubenswrapper[4926]: Trace[908235131]: ---"Objects listed" error: 12148ms (09:48:43.356) Nov 25 09:48:43 crc kubenswrapper[4926]: Trace[908235131]: [12.149027142s] [12.149027142s] END Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.356903 4926 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.400044 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49784->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.400129 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:49784->192.168.126.11:17697: read: connection reset by peer" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.565867 4926 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.565985 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.938528 4926 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.946988 4926 apiserver.go:52] "Watching apiserver" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.950349 4926 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.950811 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.951287 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.951618 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:43 crc kubenswrapper[4926]: E1125 09:48:43.951709 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.951798 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:43 crc kubenswrapper[4926]: E1125 09:48:43.951837 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.951884 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:43 crc kubenswrapper[4926]: E1125 09:48:43.951916 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.951989 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.952411 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.953697 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.954465 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.955475 4926 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.958631 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959285 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959494 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959862 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959900 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959928 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959965 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959986 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.959989 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960010 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960028 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960065 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960087 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960109 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960126 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960142 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960159 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960175 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960191 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960207 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960244 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960265 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960283 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960301 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960441 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960301 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960572 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960587 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960598 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960626 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960645 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960666 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960684 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960700 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960730 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960754 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960782 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960810 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960831 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960916 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960961 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960980 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961000 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961017 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961034 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961052 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961067 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961084 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961099 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961115 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961152 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961170 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961186 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961201 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961217 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961238 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961263 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961299 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961314 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961333 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961352 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961373 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961415 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961433 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961451 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961469 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961488 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961510 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961549 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961567 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961585 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961602 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961618 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961639 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961656 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961691 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961710 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961733 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961756 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961778 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961800 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961823 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961845 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961860 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961875 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961893 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961908 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961924 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961958 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961976 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961995 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962014 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962032 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962049 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962072 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962093 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962120 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962146 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962170 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962189 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962207 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962226 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962243 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962263 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962281 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962300 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962318 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962335 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962353 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962371 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962389 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962408 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962429 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962446 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962464 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962483 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962499 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962516 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962533 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962552 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962570 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962586 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962603 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962623 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962643 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962665 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962706 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962726 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962751 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962773 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962798 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962823 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962840 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962858 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962878 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962898 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962919 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962953 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962974 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963019 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963037 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963063 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963081 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963103 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963134 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963151 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963170 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963188 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963207 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963227 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963244 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963262 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963280 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963297 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963316 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963335 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963356 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963375 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963395 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963418 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963756 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963780 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963801 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963821 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963838 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963857 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963876 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963955 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963976 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.963997 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964017 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964038 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964060 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964082 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964126 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964156 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964184 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964206 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964232 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964258 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964275 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964295 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964321 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966009 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966069 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966121 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966387 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966423 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966448 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966469 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966599 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966727 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966953 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967064 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967188 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967340 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967367 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967460 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967719 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967981 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968119 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968246 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968282 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968340 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968367 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968459 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968483 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968592 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968780 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.968808 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.970248 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960753 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.960898 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961685 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.961958 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962110 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.962245 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964069 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964126 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.964282 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966120 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966460 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.966764 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.967412 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.969026 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.970280 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.970360 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.970391 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.994432 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.995359 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.995763 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:43 crc kubenswrapper[4926]: I1125 09:48:43.995993 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.007885 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.008534 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.008669 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.008812 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.009159 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.009410 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.009470 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.009762 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.009764 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010103 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010190 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010371 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010435 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010601 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010794 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.010882 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.011090 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.011191 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.011336 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.022760 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.023448 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.023531 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.023814 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.027729 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.029880 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.030225 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.031096 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.031182 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.031631 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.032129 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.032792 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.033114 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.033161 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.033302 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.033509 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.033836 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.033999 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.034110 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.034449 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.034649 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.034883 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.035117 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.036618 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.037105 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.037334 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.037660 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.041060 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.041282 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.042035 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.042035 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.042267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.042410 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.046814 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.049054 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.046296 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.050684 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.051520 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.051775 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052010 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052041 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052038 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052133 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052153 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.052563 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:44.55250944 +0000 UTC m=+21.236465867 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052555 4926 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.052831 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.053175 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.053265 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.054929 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.055906 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.056035 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.057211 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.057209 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.057617 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.058184 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.058765 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.059726 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.059840 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.059921 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.059993 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.060552 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.061094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.061401 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.062057 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:44.562024515 +0000 UTC m=+21.245980942 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.064014 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.064356 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.067395 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.069094 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.069371 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.069559 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.069881 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070188 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070306 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070643 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070658 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070779 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.070899 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.071031 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.071063 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.071299 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.071420 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.071604 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.072233 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.072624 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.072770 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.074409 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.074834 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.074983 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.075087 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.075332 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.075558 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.075590 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.075699 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.077128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.077153 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.077379 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.072536 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.078538 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.077038 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.079372 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.080158 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.080501 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.080901 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.080594 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.080540 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.081824 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.082394 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.082536 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.082593 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.083252 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.083294 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.083419 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.085236 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.085410 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.085581 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.085858 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.087494 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.087758 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.088904 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.089850 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.090144 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.090201 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.090397 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.090611 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.090925 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.093315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.099535 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.094767 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.099637 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.099670 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.099200 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.099852 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.099904 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.092473 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.092490 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.093387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.094211 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.094507 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:48:44.594350533 +0000 UTC m=+21.278306960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.094599 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.094953 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.095550 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.097310 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.097379 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.097415 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.097915 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.098419 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100199 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.099015 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100223 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.099071 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100291 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100316 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100356 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100375 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100391 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100517 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100595 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100687 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100703 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100719 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100746 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100761 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100776 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100797 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100816 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100833 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100848 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100865 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100881 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100897 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100913 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100928 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100965 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100984 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.100979 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101002 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101020 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101039 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101089 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101120 4926 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101143 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101203 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.101216 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:44.60118629 +0000 UTC m=+21.285142927 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101477 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101611 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.101671 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.108733 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.108794 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.110189 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.110906 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.111241 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.116225 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.118192 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.119730 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.121864 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.123733 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.125036 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.125041 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.125083 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.125101 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.125194 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:44.625166363 +0000 UTC m=+21.309122790 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.125376 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.135473 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.136515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.137054 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.137570 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.137922 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.138548 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.140370 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.140599 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.142274 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be" exitCode=255 Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.145083 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.150453 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.152694 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.153032 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.154302 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.155062 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.155894 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.157116 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.174222 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.176210 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.177869 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.179127 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.179672 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.182404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be"} Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.182701 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.199687 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.202426 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.203273 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211536 4926 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211584 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211600 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211616 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211633 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211647 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211661 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211674 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211687 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211699 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211712 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211724 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211738 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211750 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211762 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211773 4926 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211786 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211799 4926 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211811 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211823 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211835 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211845 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211855 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211864 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211872 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211881 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211889 4926 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211899 4926 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211908 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211917 4926 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211926 4926 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211951 4926 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211961 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211970 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211980 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.211990 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212000 4926 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212010 4926 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212020 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212029 4926 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212037 4926 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212048 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212057 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212066 4926 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212075 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212084 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212093 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212101 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212111 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212120 4926 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212128 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212137 4926 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212146 4926 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212155 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212166 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212178 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212189 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212200 4926 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212210 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212218 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212228 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212237 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212247 4926 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212290 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212300 4926 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212330 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212339 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212347 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212356 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212365 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.212373 4926 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.213054 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.213785 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214074 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214675 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214714 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214726 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214736 4926 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214944 4926 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214973 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214984 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.214995 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215006 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215015 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215025 4926 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215035 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215046 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215056 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215067 4926 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215076 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215086 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215096 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215106 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215115 4926 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215128 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215138 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215147 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215157 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215166 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215175 4926 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215185 4926 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215196 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215204 4926 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215213 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215228 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215237 4926 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215246 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215255 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215265 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215274 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215285 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215296 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215306 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215316 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215325 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215334 4926 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215343 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215352 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215360 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215369 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215378 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215387 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215396 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215407 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215419 4926 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215431 4926 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215442 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215451 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215460 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215470 4926 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215480 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215488 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215499 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215508 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215517 4926 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215526 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215539 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215549 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215559 4926 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215568 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215579 4926 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215589 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215598 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215608 4926 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215618 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215627 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215637 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215646 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215656 4926 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215665 4926 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215674 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215682 4926 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215691 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215702 4926 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215711 4926 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215722 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215732 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215740 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215749 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215758 4926 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215767 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215776 4926 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215786 4926 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.215854 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.236381 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.238618 4926 scope.go:117] "RemoveContainer" containerID="52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.239486 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.255227 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.288770 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.309690 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.319149 4926 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.319185 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.324181 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.339281 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.352586 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.363883 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.374311 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.380002 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.384502 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.403446 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: W1125 09:48:44.406516 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-734afbc31a3a0376c0b369896290448e518254000ff6486a194f10f88c503fe1 WatchSource:0}: Error finding container 734afbc31a3a0376c0b369896290448e518254000ff6486a194f10f88c503fe1: Status 404 returned error can't find the container with id 734afbc31a3a0376c0b369896290448e518254000ff6486a194f10f88c503fe1 Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.416842 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.418138 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.429065 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.434685 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.490188 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-58wwc"] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.493145 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.496427 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.500167 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.500212 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.516864 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.533579 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.557348 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.564615 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.575559 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.588137 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.618431 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.621978 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.622076 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.622108 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrmp6\" (UniqueName: \"kubernetes.io/projected/a918afbd-3489-49de-a796-b76a7d73748c-kube-api-access-hrmp6\") pod \"node-resolver-58wwc\" (UID: \"a918afbd-3489-49de-a796-b76a7d73748c\") " pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.622134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a918afbd-3489-49de-a796-b76a7d73748c-hosts-file\") pod \"node-resolver-58wwc\" (UID: \"a918afbd-3489-49de-a796-b76a7d73748c\") " pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.622153 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.622169 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622254 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622316 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:45.622300936 +0000 UTC m=+22.306257353 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622701 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:48:45.622691213 +0000 UTC m=+22.306647640 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622772 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622797 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:45.622790735 +0000 UTC m=+22.306747162 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622878 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622891 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622903 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.622973 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:45.622965538 +0000 UTC m=+22.306921965 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.633716 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.723527 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrmp6\" (UniqueName: \"kubernetes.io/projected/a918afbd-3489-49de-a796-b76a7d73748c-kube-api-access-hrmp6\") pod \"node-resolver-58wwc\" (UID: \"a918afbd-3489-49de-a796-b76a7d73748c\") " pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.723881 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a918afbd-3489-49de-a796-b76a7d73748c-hosts-file\") pod \"node-resolver-58wwc\" (UID: \"a918afbd-3489-49de-a796-b76a7d73748c\") " pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.724127 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.724056 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/a918afbd-3489-49de-a796-b76a7d73748c-hosts-file\") pod \"node-resolver-58wwc\" (UID: \"a918afbd-3489-49de-a796-b76a7d73748c\") " pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.724292 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.724338 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.724352 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: E1125 09:48:44.724432 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:45.724408064 +0000 UTC m=+22.408364491 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.751804 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrmp6\" (UniqueName: \"kubernetes.io/projected/a918afbd-3489-49de-a796-b76a7d73748c-kube-api-access-hrmp6\") pod \"node-resolver-58wwc\" (UID: \"a918afbd-3489-49de-a796-b76a7d73748c\") " pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.813354 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-58wwc" Nov 25 09:48:44 crc kubenswrapper[4926]: W1125 09:48:44.824702 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda918afbd_3489_49de_a796_b76a7d73748c.slice/crio-2122cfa6cfb3e93dc691dde3d7f78032c91111304c5aa2b786777cd6a3765efc WatchSource:0}: Error finding container 2122cfa6cfb3e93dc691dde3d7f78032c91111304c5aa2b786777cd6a3765efc: Status 404 returned error can't find the container with id 2122cfa6cfb3e93dc691dde3d7f78032c91111304c5aa2b786777cd6a3765efc Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.871039 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-l78n4"] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.871739 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.874061 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.874672 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.874683 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.874889 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.874927 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.877075 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vtrls"] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.878050 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.879825 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-b82dg"] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.880263 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4pcsz"] Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.880863 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.880984 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.881339 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.881960 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.883496 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.883989 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.884064 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.884242 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.885923 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.886165 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.886205 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.886242 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.886317 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.886503 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.886822 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.887106 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.888978 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.907083 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.922629 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927379 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-cni-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927432 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-config\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927463 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-os-release\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927490 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovn-node-metrics-cert\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927515 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtqrw\" (UniqueName: \"kubernetes.io/projected/9a4d98c7-0970-4ab3-86a6-40c61350f745-kube-api-access-xtqrw\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927540 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-os-release\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927583 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7142d4cf-9f77-4d00-be33-0198a4cb84d1-mcd-auth-proxy-config\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927605 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-log-socket\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927626 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-socket-dir-parent\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927647 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-cni-bin\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927667 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-ovn\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927688 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-netd\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927709 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-kubelet\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927731 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-system-cni-dir\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927775 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-node-log\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927796 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-netns\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927853 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-etc-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927873 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-ovn-kubernetes\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927898 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d167a938-f26f-449c-9ee6-bd7247ac68e4-cni-binary-copy\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927919 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-systemd-units\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927964 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-conf-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.927985 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-script-lib\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-bin\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928024 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-daemon-config\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928049 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-multus-certs\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928068 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-netns\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928090 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-systemd\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928110 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rz8v\" (UniqueName: \"kubernetes.io/projected/7142d4cf-9f77-4d00-be33-0198a4cb84d1-kube-api-access-8rz8v\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928132 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cab27fcb-cde4-4879-8fad-79951474d7b3-cni-binary-copy\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928373 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928459 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-k8s-cni-cncf-io\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928511 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-etc-kubernetes\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928542 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-var-lib-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928565 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6mn8\" (UniqueName: \"kubernetes.io/projected/cab27fcb-cde4-4879-8fad-79951474d7b3-kube-api-access-h6mn8\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqmcs\" (UniqueName: \"kubernetes.io/projected/d167a938-f26f-449c-9ee6-bd7247ac68e4-kube-api-access-pqmcs\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-system-cni-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928734 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-hostroot\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928822 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928888 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7142d4cf-9f77-4d00-be33-0198a4cb84d1-rootfs\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928909 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-cnibin\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928925 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-cni-multus\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928955 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7142d4cf-9f77-4d00-be33-0198a4cb84d1-proxy-tls\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928973 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-cnibin\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.928993 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d167a938-f26f-449c-9ee6-bd7247ac68e4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.929048 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-kubelet\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.929067 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-slash\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.929083 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-env-overrides\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.957487 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.976309 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.992131 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:44 crc kubenswrapper[4926]: I1125 09:48:44.997354 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.003675 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.006762 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.021770 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030372 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030431 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-var-lib-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030453 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-k8s-cni-cncf-io\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030481 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-etc-kubernetes\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030557 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-k8s-cni-cncf-io\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-var-lib-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030701 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-system-cni-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030725 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6mn8\" (UniqueName: \"kubernetes.io/projected/cab27fcb-cde4-4879-8fad-79951474d7b3-kube-api-access-h6mn8\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqmcs\" (UniqueName: \"kubernetes.io/projected/d167a938-f26f-449c-9ee6-bd7247ac68e4-kube-api-access-pqmcs\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030665 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-etc-kubernetes\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030785 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-cnibin\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-cni-multus\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030885 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-hostroot\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030884 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-system-cni-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030911 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030968 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7142d4cf-9f77-4d00-be33-0198a4cb84d1-rootfs\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.030993 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7142d4cf-9f77-4d00-be33-0198a4cb84d1-proxy-tls\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031020 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-cnibin\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031043 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d167a938-f26f-449c-9ee6-bd7247ac68e4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031105 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-slash\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031126 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-env-overrides\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031140 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7142d4cf-9f77-4d00-be33-0198a4cb84d1-rootfs\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031151 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-kubelet\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031175 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-cni-multus\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031176 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-cni-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031211 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-config\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtqrw\" (UniqueName: \"kubernetes.io/projected/9a4d98c7-0970-4ab3-86a6-40c61350f745-kube-api-access-xtqrw\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031233 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-cni-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-os-release\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031267 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovn-node-metrics-cert\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031295 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7142d4cf-9f77-4d00-be33-0198a4cb84d1-mcd-auth-proxy-config\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031313 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-os-release\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031331 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-cni-bin\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031349 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-ovn\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031358 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-slash\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031366 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-log-socket\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031387 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-socket-dir-parent\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031398 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-cnibin\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-system-cni-dir\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031424 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-node-log\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031440 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-netd\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031458 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-kubelet\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031478 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031493 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-etc-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-netns\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031526 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d167a938-f26f-449c-9ee6-bd7247ac68e4-cni-binary-copy\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-systemd-units\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031560 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-ovn-kubernetes\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031580 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-conf-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031598 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-script-lib\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-daemon-config\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031648 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-bin\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031669 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rz8v\" (UniqueName: \"kubernetes.io/projected/7142d4cf-9f77-4d00-be33-0198a4cb84d1-kube-api-access-8rz8v\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031689 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cab27fcb-cde4-4879-8fad-79951474d7b3-cni-binary-copy\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-multus-certs\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-netns\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-systemd\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031814 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-systemd\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031851 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-hostroot\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.031884 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.032060 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d167a938-f26f-449c-9ee6-bd7247ac68e4-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.032693 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.032737 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-config\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.032792 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-etc-openvswitch\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.032828 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-netns\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.033203 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-os-release\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.033387 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d167a938-f26f-449c-9ee6-bd7247ac68e4-cni-binary-copy\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.033450 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-systemd-units\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.033491 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-ovn-kubernetes\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.033529 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-conf-dir\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.033719 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-log-socket\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.034125 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-script-lib\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.034798 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7142d4cf-9f77-4d00-be33-0198a4cb84d1-proxy-tls\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.034819 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7142d4cf-9f77-4d00-be33-0198a4cb84d1-mcd-auth-proxy-config\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.034914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-daemon-config\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035166 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-os-release\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035201 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-node-log\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035306 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-ovn\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035267 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-multus-socket-dir-parent\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035280 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-cni-bin\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-netd\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035366 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-env-overrides\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035312 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d167a938-f26f-449c-9ee6-bd7247ac68e4-system-cni-dir\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035246 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-cnibin\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035653 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-kubelet\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035751 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-var-lib-kubelet\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035812 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/cab27fcb-cde4-4879-8fad-79951474d7b3-host-run-multus-certs\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.035852 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-netns\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.036036 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/cab27fcb-cde4-4879-8fad-79951474d7b3-cni-binary-copy\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.036109 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-bin\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.036395 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovn-node-metrics-cert\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.036807 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.051507 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqmcs\" (UniqueName: \"kubernetes.io/projected/d167a938-f26f-449c-9ee6-bd7247ac68e4-kube-api-access-pqmcs\") pod \"multus-additional-cni-plugins-vtrls\" (UID: \"d167a938-f26f-449c-9ee6-bd7247ac68e4\") " pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.052516 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6mn8\" (UniqueName: \"kubernetes.io/projected/cab27fcb-cde4-4879-8fad-79951474d7b3-kube-api-access-h6mn8\") pod \"multus-l78n4\" (UID: \"cab27fcb-cde4-4879-8fad-79951474d7b3\") " pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.052605 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtqrw\" (UniqueName: \"kubernetes.io/projected/9a4d98c7-0970-4ab3-86a6-40c61350f745-kube-api-access-xtqrw\") pod \"ovnkube-node-4pcsz\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.059556 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.059702 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rz8v\" (UniqueName: \"kubernetes.io/projected/7142d4cf-9f77-4d00-be33-0198a4cb84d1-kube-api-access-8rz8v\") pod \"machine-config-daemon-b82dg\" (UID: \"7142d4cf-9f77-4d00-be33-0198a4cb84d1\") " pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.075836 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.090430 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.120450 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.132958 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.147465 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.147540 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.147598 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2253bcbb8d617d144ff3bb4dd5e83d801501cb18fb881a9097f83513f95305ef"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.149070 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.149128 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.149210 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"734afbc31a3a0376c0b369896290448e518254000ff6486a194f10f88c503fe1"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.150756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-58wwc" event={"ID":"a918afbd-3489-49de-a796-b76a7d73748c","Type":"ContainerStarted","Data":"d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.150807 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-58wwc" event={"ID":"a918afbd-3489-49de-a796-b76a7d73748c","Type":"ContainerStarted","Data":"2122cfa6cfb3e93dc691dde3d7f78032c91111304c5aa2b786777cd6a3765efc"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.151982 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"4ff3eee43584b548545f9d55682afc2eac9f7681dc357ce2d2903f2569497b44"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.154360 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.156650 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530"} Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.162861 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.176992 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.186279 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.186701 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-l78n4" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.198497 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vtrls" Nov 25 09:48:45 crc kubenswrapper[4926]: W1125 09:48:45.199652 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcab27fcb_cde4_4879_8fad_79951474d7b3.slice/crio-119dbfb13b3953583a4e4a7db9e9ba1c427d2abdf29fef7feeb460027cd91bea WatchSource:0}: Error finding container 119dbfb13b3953583a4e4a7db9e9ba1c427d2abdf29fef7feeb460027cd91bea: Status 404 returned error can't find the container with id 119dbfb13b3953583a4e4a7db9e9ba1c427d2abdf29fef7feeb460027cd91bea Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.199869 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.207864 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:48:45 crc kubenswrapper[4926]: W1125 09:48:45.210619 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd167a938_f26f_449c_9ee6_bd7247ac68e4.slice/crio-964f54229eba24bbb569cc6347d57bbd9b09d5728138ebe2a12947d80a7395fe WatchSource:0}: Error finding container 964f54229eba24bbb569cc6347d57bbd9b09d5728138ebe2a12947d80a7395fe: Status 404 returned error can't find the container with id 964f54229eba24bbb569cc6347d57bbd9b09d5728138ebe2a12947d80a7395fe Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.213813 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.228680 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: W1125 09:48:45.235830 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7142d4cf_9f77_4d00_be33_0198a4cb84d1.slice/crio-b7faa6b604e284296fec24287792c4cdd32629838ea86858366d81c4882f8bc4 WatchSource:0}: Error finding container b7faa6b604e284296fec24287792c4cdd32629838ea86858366d81c4882f8bc4: Status 404 returned error can't find the container with id b7faa6b604e284296fec24287792c4cdd32629838ea86858366d81c4882f8bc4 Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.235880 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.249568 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: W1125 09:48:45.257128 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a4d98c7_0970_4ab3_86a6_40c61350f745.slice/crio-0e11357508f6dabaac0c0c3c16c6522bb4481cb60be7fdd562912280fd2bad02 WatchSource:0}: Error finding container 0e11357508f6dabaac0c0c3c16c6522bb4481cb60be7fdd562912280fd2bad02: Status 404 returned error can't find the container with id 0e11357508f6dabaac0c0c3c16c6522bb4481cb60be7fdd562912280fd2bad02 Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.270758 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.291141 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.316826 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.331511 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.368271 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.382777 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.396313 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.414409 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.427795 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.445644 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.462571 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.478906 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.498873 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.512819 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:45Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.640013 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.640191 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640268 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:48:47.640227188 +0000 UTC m=+24.324183615 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.640500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640529 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.640568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640573 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:47.640565584 +0000 UTC m=+24.324522011 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640425 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640632 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640646 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640675 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:47.640666766 +0000 UTC m=+24.324623193 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640733 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.640758 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:47.640750658 +0000 UTC m=+24.324707085 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: I1125 09:48:45.741840 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.742210 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.742258 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.742274 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:45 crc kubenswrapper[4926]: E1125 09:48:45.742368 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:47.742336136 +0000 UTC m=+24.426292753 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.012737 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:46 crc kubenswrapper[4926]: E1125 09:48:46.012876 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.013305 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:46 crc kubenswrapper[4926]: E1125 09:48:46.013379 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.013423 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:46 crc kubenswrapper[4926]: E1125 09:48:46.013466 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.017661 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.018490 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.020156 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.020845 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.021979 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.022641 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.023417 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.024527 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.025328 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.026458 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.027153 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.028589 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.029352 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.029874 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.030864 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.031388 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.032419 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.033197 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.034286 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.034772 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.035317 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.036199 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.036874 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.037857 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.038501 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.039499 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.039964 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.040908 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.041504 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.042044 4926 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.042153 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.044330 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.044855 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.046139 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.047743 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.048523 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.049521 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.050385 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.051533 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.052056 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.161165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.161221 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.161230 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"b7faa6b604e284296fec24287792c4cdd32629838ea86858366d81c4882f8bc4"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.162964 4926 generic.go:334] "Generic (PLEG): container finished" podID="d167a938-f26f-449c-9ee6-bd7247ac68e4" containerID="30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15" exitCode=0 Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.163015 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerDied","Data":"30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.163036 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerStarted","Data":"964f54229eba24bbb569cc6347d57bbd9b09d5728138ebe2a12947d80a7395fe"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.164525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerStarted","Data":"53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.164643 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerStarted","Data":"119dbfb13b3953583a4e4a7db9e9ba1c427d2abdf29fef7feeb460027cd91bea"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.166891 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" exitCode=0 Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.166996 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.167030 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"0e11357508f6dabaac0c0c3c16c6522bb4481cb60be7fdd562912280fd2bad02"} Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.167345 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.181456 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.199052 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.223827 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.255233 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.276270 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.290808 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.316016 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.351506 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.424336 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.449034 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.467413 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.485367 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.510164 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.534271 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.560430 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.578178 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.595893 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.616205 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.633692 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.651287 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.672206 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.685632 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.697618 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.712399 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.726784 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:46 crc kubenswrapper[4926]: I1125 09:48:46.739087 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:46Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.175680 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.176054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.176069 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.179095 4926 generic.go:334] "Generic (PLEG): container finished" podID="d167a938-f26f-449c-9ee6-bd7247ac68e4" containerID="ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07" exitCode=0 Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.179178 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerDied","Data":"ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07"} Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.198246 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.214310 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.235126 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.255604 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.270512 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.285591 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.299802 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.323173 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.338556 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.349093 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.365065 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.377662 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.393856 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.408724 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.423413 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.437185 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.449515 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.462328 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.474660 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.485715 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.496679 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.510977 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.523995 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.536671 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.550700 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.589142 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-pfggm"] Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.589409 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.589624 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.593948 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.594395 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.594494 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.594741 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.615384 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.638023 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.655336 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.662823 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.662929 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrkks\" (UniqueName: \"kubernetes.io/projected/de954742-8a79-473e-bcb4-537799771043-kube-api-access-zrkks\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.662982 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663048 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:48:51.663007623 +0000 UTC m=+28.346964100 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663102 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.663127 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663155 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:51.663139885 +0000 UTC m=+28.347096312 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.663204 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de954742-8a79-473e-bcb4-537799771043-host\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.663226 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/de954742-8a79-473e-bcb4-537799771043-serviceca\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.663264 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663334 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663408 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:51.66339461 +0000 UTC m=+28.347351047 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663413 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663443 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663472 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.663522 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:51.663510612 +0000 UTC m=+28.347467229 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.667968 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.684528 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.727397 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.764691 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/de954742-8a79-473e-bcb4-537799771043-serviceca\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.764747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de954742-8a79-473e-bcb4-537799771043-host\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.764791 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.764819 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrkks\" (UniqueName: \"kubernetes.io/projected/de954742-8a79-473e-bcb4-537799771043-kube-api-access-zrkks\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.764869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/de954742-8a79-473e-bcb4-537799771043-host\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.765055 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.765096 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.765112 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:47 crc kubenswrapper[4926]: E1125 09:48:47.765186 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:51.765166702 +0000 UTC m=+28.449123129 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.766148 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/de954742-8a79-473e-bcb4-537799771043-serviceca\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.769059 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.777931 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.789462 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.797176 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrkks\" (UniqueName: \"kubernetes.io/projected/de954742-8a79-473e-bcb4-537799771043-kube-api-access-zrkks\") pod \"node-ca-pfggm\" (UID: \"de954742-8a79-473e-bcb4-537799771043\") " pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.830165 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.847166 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.887207 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.904309 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-pfggm" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.931273 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:47 crc kubenswrapper[4926]: I1125 09:48:47.967533 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:47Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.007481 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.011829 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.011888 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:48 crc kubenswrapper[4926]: E1125 09:48:48.011960 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.011888 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:48 crc kubenswrapper[4926]: E1125 09:48:48.012065 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:48 crc kubenswrapper[4926]: E1125 09:48:48.012090 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.048200 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.092884 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.129050 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.168102 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.188092 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pfggm" event={"ID":"de954742-8a79-473e-bcb4-537799771043","Type":"ContainerStarted","Data":"a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.188145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-pfggm" event={"ID":"de954742-8a79-473e-bcb4-537799771043","Type":"ContainerStarted","Data":"20877d77e60ac252463efec0b144d912d3eb20e28d3797bd643ac21de2cea642"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.189441 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.192527 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.192557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.192571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.194346 4926 generic.go:334] "Generic (PLEG): container finished" podID="d167a938-f26f-449c-9ee6-bd7247ac68e4" containerID="105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305" exitCode=0 Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.194441 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerDied","Data":"105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305"} Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.208081 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.247235 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.288293 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.333595 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.367326 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.408797 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.452505 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.493933 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.528796 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.568516 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.608744 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.648317 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.687468 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.730176 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.773653 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.810258 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.848365 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.889844 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.926318 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:48 crc kubenswrapper[4926]: I1125 09:48:48.968018 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:48Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.009578 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.048080 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.086211 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.126786 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.164416 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.201401 4926 generic.go:334] "Generic (PLEG): container finished" podID="d167a938-f26f-449c-9ee6-bd7247ac68e4" containerID="c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0" exitCode=0 Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.201457 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerDied","Data":"c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0"} Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.217849 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.250979 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.287895 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.335780 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.367914 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.409001 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.449492 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.493241 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.532884 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.570166 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.608064 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.648666 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.688105 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.729853 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.755089 4926 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.757103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.757146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.757158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.757308 4926 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.768627 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.821557 4926 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.821879 4926 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.823276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.823334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.823348 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.823371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.823386 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:49Z","lastTransitionTime":"2025-11-25T09:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:49 crc kubenswrapper[4926]: E1125 09:48:49.837769 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.841997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.842058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.842074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.842094 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.842110 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:49Z","lastTransitionTime":"2025-11-25T09:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.848453 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: E1125 09:48:49.854853 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.859431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.859475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.859491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.859514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.859530 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:49Z","lastTransitionTime":"2025-11-25T09:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:49 crc kubenswrapper[4926]: E1125 09:48:49.872223 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.876207 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.876250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.876264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.876282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.876293 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:49Z","lastTransitionTime":"2025-11-25T09:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.890235 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: E1125 09:48:49.891462 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.895888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.895954 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.895966 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.895987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.895999 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:49Z","lastTransitionTime":"2025-11-25T09:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:49 crc kubenswrapper[4926]: E1125 09:48:49.907386 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:49 crc kubenswrapper[4926]: E1125 09:48:49.907548 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.909395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.909429 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.909440 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.909461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.909472 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:49Z","lastTransitionTime":"2025-11-25T09:48:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:49 crc kubenswrapper[4926]: I1125 09:48:49.930745 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:49Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.011387 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.011459 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.011554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:50 crc kubenswrapper[4926]: E1125 09:48:50.011546 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:50 crc kubenswrapper[4926]: E1125 09:48:50.011709 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:50 crc kubenswrapper[4926]: E1125 09:48:50.011770 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.012983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.013063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.013076 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.013095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.013110 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.115648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.115689 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.115698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.115716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.115728 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.209174 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.211748 4926 generic.go:334] "Generic (PLEG): container finished" podID="d167a938-f26f-449c-9ee6-bd7247ac68e4" containerID="21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da" exitCode=0 Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.211787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerDied","Data":"21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.217927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.217973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.217982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.217999 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.218011 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.231886 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.256295 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.269184 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.283071 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.296727 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.312640 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.320159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.320203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.320213 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.320231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.320268 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.327849 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.341088 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.354617 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.365394 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.378342 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.410311 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.423084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.423235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.423307 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.423382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.423447 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.449875 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.488487 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.526600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.526646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.526658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.526686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.526699 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.534405 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:50Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.630187 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.630240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.630250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.630270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.630282 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.732634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.732686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.732696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.732714 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.732725 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.835395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.835449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.835467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.835490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.835504 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.938337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.938386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.938405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.938424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:50 crc kubenswrapper[4926]: I1125 09:48:50.938435 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:50Z","lastTransitionTime":"2025-11-25T09:48:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.041647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.041977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.041993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.042013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.042026 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.144432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.144481 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.144493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.144511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.144522 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.218544 4926 generic.go:334] "Generic (PLEG): container finished" podID="d167a938-f26f-449c-9ee6-bd7247ac68e4" containerID="42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833" exitCode=0 Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.218605 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerDied","Data":"42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.240758 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.247748 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.247796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.247805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.247823 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.247833 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.260431 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.275328 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.293903 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.313191 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.327312 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.340999 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.350711 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.350769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.350781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.350806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.350819 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.355338 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.365360 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.374139 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.391698 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.403598 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.415307 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.425450 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.439628 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.453626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.453676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.453688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.453706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.453716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.556602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.556644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.556664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.556684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.556696 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.665234 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.665295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.665306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.665330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.665344 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.706504 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.706599 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.706632 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.706658 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706745 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:48:59.706709697 +0000 UTC m=+36.390666124 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706765 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706827 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:59.706817079 +0000 UTC m=+36.390773506 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706850 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706869 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706881 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.706968 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:59.706924371 +0000 UTC m=+36.390880978 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.707015 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.707043 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:59.707035103 +0000 UTC m=+36.390991730 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.768058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.768100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.768110 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.768129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.768141 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.808475 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.808779 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.808844 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.808861 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:51 crc kubenswrapper[4926]: E1125 09:48:51.808965 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:48:59.808926827 +0000 UTC m=+36.492883424 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.870997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.871032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.871041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.871058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.871067 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.974090 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.974139 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.974152 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.974171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:51 crc kubenswrapper[4926]: I1125 09:48:51.974194 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:51Z","lastTransitionTime":"2025-11-25T09:48:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.011732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.011762 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.011732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:52 crc kubenswrapper[4926]: E1125 09:48:52.011904 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:52 crc kubenswrapper[4926]: E1125 09:48:52.011952 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:52 crc kubenswrapper[4926]: E1125 09:48:52.012022 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.077131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.077188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.077199 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.077228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.077242 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.180606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.180662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.180674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.180698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.180716 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.227273 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.227674 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.227734 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.235022 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" event={"ID":"d167a938-f26f-449c-9ee6-bd7247ac68e4","Type":"ContainerStarted","Data":"b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.245403 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.258513 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.258987 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.260052 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.278070 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.283200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.283537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.283628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.283710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.283803 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.291043 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.302494 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.315477 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.326868 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.338323 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.348305 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.359582 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.381867 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.386970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.387035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.387050 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.387073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.387087 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.396457 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.410490 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.425439 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.449644 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.473291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.487498 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.489452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.489494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.489507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.489527 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.489537 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.500574 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.517359 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.537824 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.555718 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.572578 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.589017 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.592960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.593010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.593021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.593038 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.593049 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.604803 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.619048 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.634634 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.647925 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.662727 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.674140 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.688677 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:52Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.696072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.696122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.696133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.696153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.696167 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.798404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.798444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.798455 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.798473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.798488 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.901331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.901407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.901423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.901444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:52 crc kubenswrapper[4926]: I1125 09:48:52.901457 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:52Z","lastTransitionTime":"2025-11-25T09:48:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.005064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.005112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.005124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.005166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.005177 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.109249 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.109653 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.109744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.109835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.109919 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.213252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.213297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.213308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.213330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.213342 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.238786 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.315878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.315963 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.315983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.316009 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.316026 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.419420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.419472 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.419485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.419505 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.419519 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.522915 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.522974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.522987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.523007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.523021 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.626109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.626160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.626173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.626196 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.626216 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.729367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.729412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.729432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.729453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.729467 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.832273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.832320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.832332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.832353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.832364 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.934910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.934973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.934983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.935001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:53 crc kubenswrapper[4926]: I1125 09:48:53.935014 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:53Z","lastTransitionTime":"2025-11-25T09:48:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.012093 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:54 crc kubenswrapper[4926]: E1125 09:48:54.012240 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.012645 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:54 crc kubenswrapper[4926]: E1125 09:48:54.012715 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.012770 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:54 crc kubenswrapper[4926]: E1125 09:48:54.012823 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.028172 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.037717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.037771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.037784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.037802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.037815 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.042902 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.059181 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.071115 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.086039 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.099454 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.112819 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.135252 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.140432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.140472 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.140485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.140503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.140516 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.162129 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.183594 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.232614 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.242524 4926 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.243251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.243287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.243295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.243316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.243327 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.262202 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.277152 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.295922 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.318021 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.346327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.346709 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.346814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.346919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.347030 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.450171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.450203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.450212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.450228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.450239 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.553339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.553389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.553398 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.553415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.553428 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.656055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.656093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.656102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.656121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.656132 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.758931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.759035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.759047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.759069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.759083 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.862583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.862641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.862652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.862673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.862686 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.965483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.965529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.965545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.965566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:54 crc kubenswrapper[4926]: I1125 09:48:54.965582 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:54Z","lastTransitionTime":"2025-11-25T09:48:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.067949 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.068004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.068013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.068035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.068046 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.171161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.171226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.171237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.171258 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.171268 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.248611 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/0.log" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.252628 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20" exitCode=1 Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.252690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.256528 4926 scope.go:117] "RemoveContainer" containerID="6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.269778 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.275131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.275204 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.275215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.275233 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.275245 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.289267 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.310775 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.329760 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.350028 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:55Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:48:55.124247 6242 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 09:48:55.124263 6242 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 09:48:55.124293 6242 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:48:55.124336 6242 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:48:55.124347 6242 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:48:55.124360 6242 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:48:55.124378 6242 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:55.124401 6242 factory.go:656] Stopping watch factory\\\\nI1125 09:48:55.124420 6242 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:55.124441 6242 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:48:55.124454 6242 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:48:55.124460 6242 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:48:55.124466 6242 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:48:55.124472 6242 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:48:55.124484 6242 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.364686 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.378257 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.378786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.378864 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.378882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.378910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.378925 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.392218 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.411546 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.427017 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.441233 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.453689 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.463629 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.475609 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.481238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.481287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.481304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.481328 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.481344 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.494302 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:55Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.584384 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.584437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.584449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.584469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.584483 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.686926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.686999 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.687012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.687032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.687044 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.790530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.790582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.790593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.790612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.790625 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.893282 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.893326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.893337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.893358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.893370 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.996431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.996478 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.996490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.996511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:55 crc kubenswrapper[4926]: I1125 09:48:55.996524 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:55Z","lastTransitionTime":"2025-11-25T09:48:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.011819 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.011866 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.011931 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:56 crc kubenswrapper[4926]: E1125 09:48:56.012014 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:56 crc kubenswrapper[4926]: E1125 09:48:56.012137 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:56 crc kubenswrapper[4926]: E1125 09:48:56.012245 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.099996 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.100047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.100056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.100074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.100084 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.119210 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.202961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.203019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.203033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.203056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.203069 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.258006 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/0.log" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.261684 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.262258 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.275797 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.289006 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.301850 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.305405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.305434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.305445 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.305462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.305472 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.311582 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.328363 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.345621 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.360974 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.374990 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.386480 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.400275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.408542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.408597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.408610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.408655 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.408669 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.413457 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.468741 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.492591 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.511342 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:55Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:48:55.124247 6242 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 09:48:55.124263 6242 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 09:48:55.124293 6242 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:48:55.124336 6242 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:48:55.124347 6242 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:48:55.124360 6242 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:48:55.124378 6242 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:55.124401 6242 factory.go:656] Stopping watch factory\\\\nI1125 09:48:55.124420 6242 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:55.124441 6242 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:48:55.124454 6242 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:48:55.124460 6242 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:48:55.124466 6242 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:48:55.124472 6242 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:48:55.124484 6242 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.511967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.512003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.512014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.512032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.512043 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.525637 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:56Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.615134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.615199 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.615211 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.615234 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.615246 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.717579 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.717624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.717636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.717769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.717815 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.821068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.821120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.821131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.821152 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.821164 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.924389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.924437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.924452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.924471 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:56 crc kubenswrapper[4926]: I1125 09:48:56.924484 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:56Z","lastTransitionTime":"2025-11-25T09:48:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.027203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.027245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.027273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.027299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.027315 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.129974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.130019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.130031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.130052 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.130084 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.233399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.233457 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.233472 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.233495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.233507 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.267745 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/1.log" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.268465 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/0.log" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.271206 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671" exitCode=1 Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.271259 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.271309 4926 scope.go:117] "RemoveContainer" containerID="6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.272043 4926 scope.go:117] "RemoveContainer" containerID="3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671" Nov 25 09:48:57 crc kubenswrapper[4926]: E1125 09:48:57.272687 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.289852 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.311234 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:55Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:48:55.124247 6242 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 09:48:55.124263 6242 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 09:48:55.124293 6242 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:48:55.124336 6242 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:48:55.124347 6242 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:48:55.124360 6242 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:48:55.124378 6242 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:55.124401 6242 factory.go:656] Stopping watch factory\\\\nI1125 09:48:55.124420 6242 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:55.124441 6242 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:48:55.124454 6242 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:48:55.124460 6242 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:48:55.124466 6242 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:48:55.124472 6242 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:48:55.124484 6242 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.324701 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.335397 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.337129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.337178 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.337188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.337207 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.337217 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.351142 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.367520 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.381808 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.395866 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.406505 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.421061 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.434995 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.439551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.439598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.439610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.439629 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.439641 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.450150 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.463258 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.484631 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.502795 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.542870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.542909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.542920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.542957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.542969 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.618643 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l"] Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.619277 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.621130 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.621757 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.632641 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.646003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.646058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.646072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.646091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.646102 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.651990 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.665210 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.678622 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.699394 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.711341 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.725552 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.739758 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.750908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.751005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.751025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.751056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.751074 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.757967 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.771481 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/27b20208-edb8-4717-876c-ce1991660bc8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.771542 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/27b20208-edb8-4717-876c-ce1991660bc8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.771585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-544bk\" (UniqueName: \"kubernetes.io/projected/27b20208-edb8-4717-876c-ce1991660bc8-kube-api-access-544bk\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.771623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/27b20208-edb8-4717-876c-ce1991660bc8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.772003 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.792066 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.806057 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.823823 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.853778 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.854710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.854754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.854770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.854792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.854807 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.872597 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-544bk\" (UniqueName: \"kubernetes.io/projected/27b20208-edb8-4717-876c-ce1991660bc8-kube-api-access-544bk\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.872673 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/27b20208-edb8-4717-876c-ce1991660bc8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.872707 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/27b20208-edb8-4717-876c-ce1991660bc8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.872754 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/27b20208-edb8-4717-876c-ce1991660bc8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.873563 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/27b20208-edb8-4717-876c-ce1991660bc8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.873636 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/27b20208-edb8-4717-876c-ce1991660bc8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.881844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/27b20208-edb8-4717-876c-ce1991660bc8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.890461 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6da53846fb590ef149ed5f3fecfc0a96ce2016a30a8b90eb74aa19b3939d7c20\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:55Z\\\",\\\"message\\\":\\\"all/v1/apis/informers/externalversions/factory.go:140\\\\nI1125 09:48:55.124247 6242 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1125 09:48:55.124263 6242 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1125 09:48:55.124293 6242 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1125 09:48:55.124336 6242 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1125 09:48:55.124347 6242 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1125 09:48:55.124360 6242 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1125 09:48:55.124378 6242 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:55.124401 6242 factory.go:656] Stopping watch factory\\\\nI1125 09:48:55.124420 6242 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:55.124441 6242 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1125 09:48:55.124454 6242 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1125 09:48:55.124460 6242 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1125 09:48:55.124466 6242 handler.go:208] Removed *v1.Node event handler 2\\\\nI1125 09:48:55.124472 6242 handler.go:208] Removed *v1.Node event handler 7\\\\nI1125 09:48:55.124484 6242 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1125 09:48:5\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.893507 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-544bk\" (UniqueName: \"kubernetes.io/projected/27b20208-edb8-4717-876c-ce1991660bc8-kube-api-access-544bk\") pod \"ovnkube-control-plane-749d76644c-vbg2l\" (UID: \"27b20208-edb8-4717-876c-ce1991660bc8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.902105 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:57Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.931544 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" Nov 25 09:48:57 crc kubenswrapper[4926]: W1125 09:48:57.946625 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27b20208_edb8_4717_876c_ce1991660bc8.slice/crio-64c7b936e2443894222f5249058fa910f3315e16dd68cf1f00cc2620c63848b7 WatchSource:0}: Error finding container 64c7b936e2443894222f5249058fa910f3315e16dd68cf1f00cc2620c63848b7: Status 404 returned error can't find the container with id 64c7b936e2443894222f5249058fa910f3315e16dd68cf1f00cc2620c63848b7 Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.959438 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.959504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.959515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.959535 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:57 crc kubenswrapper[4926]: I1125 09:48:57.959551 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:57Z","lastTransitionTime":"2025-11-25T09:48:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.012173 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.012287 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.012287 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.012485 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.012638 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.012678 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.062794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.062836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.062848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.062868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.062880 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.165602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.165651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.165661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.165678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.165689 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.269836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.269875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.269885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.269901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.269913 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.283032 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/1.log" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.288232 4926 scope.go:117] "RemoveContainer" containerID="3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671" Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.288435 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.290501 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" event={"ID":"27b20208-edb8-4717-876c-ce1991660bc8","Type":"ContainerStarted","Data":"7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.290555 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" event={"ID":"27b20208-edb8-4717-876c-ce1991660bc8","Type":"ContainerStarted","Data":"2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.290567 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" event={"ID":"27b20208-edb8-4717-876c-ce1991660bc8","Type":"ContainerStarted","Data":"64c7b936e2443894222f5249058fa910f3315e16dd68cf1f00cc2620c63848b7"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.307248 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.325291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.339265 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.346141 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-xpn28"] Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.346828 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.346906 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.364780 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.372729 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.372800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.372814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.372836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.372868 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.385270 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.399215 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.414337 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.430073 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.444873 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.457538 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.469057 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.476608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.476647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.476659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.476678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.476689 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.481012 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.481157 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh6kd\" (UniqueName: \"kubernetes.io/projected/81f30c43-dacf-407f-adb3-bdbae866485f-kube-api-access-dh6kd\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.493725 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.510072 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.523970 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.537577 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.554062 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.567751 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.579197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.579240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.579255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.579319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.579332 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.582681 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.582731 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh6kd\" (UniqueName: \"kubernetes.io/projected/81f30c43-dacf-407f-adb3-bdbae866485f-kube-api-access-dh6kd\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.582857 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:48:58 crc kubenswrapper[4926]: E1125 09:48:58.582918 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:48:59.082902578 +0000 UTC m=+35.766859005 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.588869 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.602072 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.605644 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh6kd\" (UniqueName: \"kubernetes.io/projected/81f30c43-dacf-407f-adb3-bdbae866485f-kube-api-access-dh6kd\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.616136 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.627198 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.642876 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.654600 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.664313 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.676817 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.681230 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.681265 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.681276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.681294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.681306 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.691060 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.715585 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.728819 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.743020 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.755009 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.771404 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.784057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.784111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.784123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.784145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.784158 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.790968 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.803172 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:48:58Z is after 2025-08-24T17:21:41Z" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.886679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.886726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.886736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.886755 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.886765 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.990220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.990271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.990287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.990305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:58 crc kubenswrapper[4926]: I1125 09:48:58.990318 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:58Z","lastTransitionTime":"2025-11-25T09:48:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.087164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.087321 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.087390 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:49:00.087372257 +0000 UTC m=+36.771328694 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.092662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.092723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.092740 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.092765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.092843 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.196628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.196699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.196717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.196737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.196750 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.299651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.299706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.299724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.299749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.299766 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.402639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.402696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.402710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.402732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.402748 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.507149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.507427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.507436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.507455 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.507464 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.609630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.609673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.609683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.609717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.609731 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.714365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.714410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.714423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.714443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.714456 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.794996 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.795149 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.795183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.795225 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795321 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795379 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:15.795364668 +0000 UTC m=+52.479321095 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795752 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:49:15.795742155 +0000 UTC m=+52.479698582 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795831 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795854 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:15.795848287 +0000 UTC m=+52.479804714 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795908 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795920 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.795963 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.796003 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:15.79599227 +0000 UTC m=+52.479948697 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.816592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.816909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.817030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.817185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.817268 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.896530 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.896768 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.896813 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.896827 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:59 crc kubenswrapper[4926]: E1125 09:48:59.896909 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:15.896887435 +0000 UTC m=+52.580843862 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.920479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.920811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.921011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.921109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:48:59 crc kubenswrapper[4926]: I1125 09:48:59.921202 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:48:59Z","lastTransitionTime":"2025-11-25T09:48:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.011686 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.011812 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.011834 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.011880 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.011953 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.012012 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.012147 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.012367 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.024209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.024256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.024267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.024283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.024295 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.066046 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.066124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.066141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.066167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.066185 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.080134 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.083828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.083961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.084045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.084144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.084208 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.096896 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.098358 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.098531 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.098607 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:49:02.098590425 +0000 UTC m=+38.782546852 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.101543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.101582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.101594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.101613 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.101626 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.115584 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.119694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.119770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.119792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.119820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.119833 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.135090 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.139167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.139316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.139388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.139451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.139548 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.151166 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:00Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:00 crc kubenswrapper[4926]: E1125 09:49:00.151301 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.153357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.153409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.153421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.153441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.153452 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.256861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.256906 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.256918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.256956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.256969 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.359730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.359844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.359857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.359879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.359913 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.463252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.463300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.463309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.463328 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.463338 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.565517 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.565557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.565565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.565582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.565592 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.673462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.673553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.673567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.673589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.673603 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.776695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.776726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.776734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.776748 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.776757 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.880285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.880361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.880379 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.880408 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.880427 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.983848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.983895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.983904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.983920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:00 crc kubenswrapper[4926]: I1125 09:49:00.983931 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:00Z","lastTransitionTime":"2025-11-25T09:49:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.086802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.086851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.086862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.086882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.086893 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.189479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.189520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.189531 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.189557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.189569 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.292347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.292443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.292466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.292493 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.292510 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.396482 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.396544 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.396557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.396578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.396594 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.499687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.499731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.499743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.499765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.499783 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.602590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.602638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.602647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.602669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.602680 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.705232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.705292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.705313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.705338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.705352 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.808295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.808356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.808365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.808382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.808394 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.820665 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.838460 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.852479 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.867037 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.878627 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.891206 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.904737 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.911071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.911109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.911118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.911136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.911149 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:01Z","lastTransitionTime":"2025-11-25T09:49:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.918050 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.932429 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.945847 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.956465 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:01 crc kubenswrapper[4926]: I1125 09:49:01.996690 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.009840 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:02Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.013667 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:02 crc kubenswrapper[4926]: E1125 09:49:02.013804 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.014236 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:02 crc kubenswrapper[4926]: E1125 09:49:02.014291 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.014330 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:02 crc kubenswrapper[4926]: E1125 09:49:02.014374 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.014409 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:02 crc kubenswrapper[4926]: E1125 09:49:02.014447 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.016340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.016365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.016373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.016387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.016399 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.025075 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:02Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.038833 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:02Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.054699 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:02Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.072156 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:02Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.083620 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:02Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.119370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.119411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.119420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.119434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.119445 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.124253 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:02 crc kubenswrapper[4926]: E1125 09:49:02.124479 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:02 crc kubenswrapper[4926]: E1125 09:49:02.124605 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:49:06.12457908 +0000 UTC m=+42.808535677 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.222330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.222401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.222416 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.222452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.222471 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.326141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.326200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.326248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.326272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.326286 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.429296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.429342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.429365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.429386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.429421 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.532052 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.532095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.532114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.532135 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.532147 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.634842 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.634927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.634985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.635018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.635041 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.737988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.738032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.738043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.738069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.738087 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.840656 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.840720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.840730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.840755 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.840778 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.943638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.943683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.943695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.943715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:02 crc kubenswrapper[4926]: I1125 09:49:02.943728 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:02Z","lastTransitionTime":"2025-11-25T09:49:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.046588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.046698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.046725 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.046754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.046773 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.149251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.149306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.149319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.149340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.149353 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.252284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.252348 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.252368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.252402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.252416 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.355173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.355246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.355270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.355299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.355318 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.457691 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.457738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.457756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.457775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.457788 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.560928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.561100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.561136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.561173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.561199 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.664224 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.664285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.664297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.664317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.664332 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.767297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.767336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.767346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.767364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.767378 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.870222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.870269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.870279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.870296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.870306 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.972960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.973011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.973024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.973048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:03 crc kubenswrapper[4926]: I1125 09:49:03.973061 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:03Z","lastTransitionTime":"2025-11-25T09:49:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.011459 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.011657 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.011505 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:04 crc kubenswrapper[4926]: E1125 09:49:04.011701 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.011505 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:04 crc kubenswrapper[4926]: E1125 09:49:04.011862 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:04 crc kubenswrapper[4926]: E1125 09:49:04.012007 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:04 crc kubenswrapper[4926]: E1125 09:49:04.012096 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.025592 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.038682 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.053403 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.069208 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.079896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.079984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.080001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.080018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.080052 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.091771 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.104800 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.119151 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.131829 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.143009 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.159805 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.179304 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.183142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.183183 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.183191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.183208 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.183219 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.194886 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.207680 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.217487 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.229137 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.245818 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.258842 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.285718 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.285771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.285782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.285801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.285813 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.389071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.389149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.389168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.389196 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.389214 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.492029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.492073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.492083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.492102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.492113 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.595010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.595055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.595065 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.595084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.595094 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.698148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.698193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.698208 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.698230 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.698247 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.800926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.800995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.801008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.801027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.801040 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.904403 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.904441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.904452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.904471 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:04 crc kubenswrapper[4926]: I1125 09:49:04.904482 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:04Z","lastTransitionTime":"2025-11-25T09:49:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.006954 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.007013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.007027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.007083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.007094 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.110219 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.110555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.110675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.110775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.110857 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.214749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.214787 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.214795 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.214812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.214824 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.316777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.316822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.316835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.316852 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.316862 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.420113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.420158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.420169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.420187 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.420200 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.522169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.522219 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.522230 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.522247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.522258 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.624971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.625033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.625047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.625069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.625083 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.727701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.727768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.727782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.727804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.727817 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.830589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.830643 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.830657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.830676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.830688 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.933275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.933347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.933367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.933417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:05 crc kubenswrapper[4926]: I1125 09:49:05.933435 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:05Z","lastTransitionTime":"2025-11-25T09:49:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.012208 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.012279 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.012272 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.012226 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:06 crc kubenswrapper[4926]: E1125 09:49:06.012383 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:06 crc kubenswrapper[4926]: E1125 09:49:06.012560 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:06 crc kubenswrapper[4926]: E1125 09:49:06.012688 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:06 crc kubenswrapper[4926]: E1125 09:49:06.012824 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.036657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.036961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.037026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.037116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.037233 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.140417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.140788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.140965 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.141106 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.141231 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.170555 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:06 crc kubenswrapper[4926]: E1125 09:49:06.170714 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:06 crc kubenswrapper[4926]: E1125 09:49:06.170789 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:49:14.170769941 +0000 UTC m=+50.854726378 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.244512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.244665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.244687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.244750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.244771 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.348232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.348299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.348315 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.348344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.348361 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.451827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.451879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.451892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.451920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.451956 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.555053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.555408 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.555434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.555459 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.555472 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.658228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.658620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.658736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.658825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.658921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.762130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.762458 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.762541 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.762627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.762692 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.865993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.866052 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.866066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.866088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.866108 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.969747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.969833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.969854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.969881 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:06 crc kubenswrapper[4926]: I1125 09:49:06.969905 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:06Z","lastTransitionTime":"2025-11-25T09:49:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.073603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.073668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.073683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.073711 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.073736 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.177047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.177092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.177103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.177120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.177144 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.281404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.281456 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.281468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.281487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.281506 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.384060 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.384151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.384175 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.384209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.384233 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.487208 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.487284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.487311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.487346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.487370 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.591057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.591105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.591117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.591137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.591149 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.694238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.694289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.694302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.694321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.694787 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.797782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.797851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.797862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.797887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.797904 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.901367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.901422 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.901431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.901449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:07 crc kubenswrapper[4926]: I1125 09:49:07.901463 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:07Z","lastTransitionTime":"2025-11-25T09:49:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.004269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.004401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.004417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.004436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.004448 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.011639 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.011694 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.011639 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:08 crc kubenswrapper[4926]: E1125 09:49:08.011780 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:08 crc kubenswrapper[4926]: E1125 09:49:08.011891 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:08 crc kubenswrapper[4926]: E1125 09:49:08.012085 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.012144 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:08 crc kubenswrapper[4926]: E1125 09:49:08.012235 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.107122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.107167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.107180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.107199 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.107211 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.210235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.210295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.210310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.210328 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.210339 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.313390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.313435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.313446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.313464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.313477 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.416668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.416732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.416747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.416780 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.416797 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.519329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.519361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.519371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.519386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.519396 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.621578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.621620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.621633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.621659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.621673 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.725154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.725215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.725231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.725253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.725268 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.828109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.828177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.828194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.828221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.828238 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.931206 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.931272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.931289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.931319 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:08 crc kubenswrapper[4926]: I1125 09:49:08.931337 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:08Z","lastTransitionTime":"2025-11-25T09:49:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.034448 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.034507 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.034527 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.034553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.034568 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.137024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.137064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.137074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.137089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.137099 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.239807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.239856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.239868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.239888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.239902 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.343771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.343844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.343868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.343902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.343929 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.446724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.446792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.446830 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.446857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.446880 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.550358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.550405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.550417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.550436 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.550450 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.653994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.654037 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.654045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.654062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.654077 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.756696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.756747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.756759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.756781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.756793 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.859695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.859740 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.859752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.859769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.859788 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.962491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.962558 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.962575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.962604 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:09 crc kubenswrapper[4926]: I1125 09:49:09.962627 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:09Z","lastTransitionTime":"2025-11-25T09:49:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.011996 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.012084 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.012164 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.011996 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.012031 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.012256 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.012354 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.012440 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.066237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.066312 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.066332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.066360 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.066380 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.169988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.170053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.170069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.170091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.170105 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.217331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.217369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.217378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.217394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.217403 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.230163 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:10Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.234593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.234626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.234636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.234659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.234670 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.245263 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:10Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.248683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.248758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.248770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.248790 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.248805 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.260974 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:10Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.264478 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.264509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.264521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.264538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.264549 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.276098 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:10Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.280051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.280098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.280108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.280131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.280144 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.293158 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:10Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:10 crc kubenswrapper[4926]: E1125 09:49:10.293287 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.294993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.295031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.295041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.295059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.295073 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.397166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.397200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.397210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.397227 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.397237 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.500478 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.501279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.501326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.501353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.501375 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.604316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.604365 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.604374 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.604393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.604405 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.706565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.706599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.706608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.706622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.706634 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.809331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.809371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.809381 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.809397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.809407 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.913470 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.913525 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.913538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.913560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:10 crc kubenswrapper[4926]: I1125 09:49:10.913575 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:10Z","lastTransitionTime":"2025-11-25T09:49:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.012893 4926 scope.go:117] "RemoveContainer" containerID="3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.016487 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.016559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.016574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.016598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.016613 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.120148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.120750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.120775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.120809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.120837 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.223695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.223731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.223740 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.223758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.223768 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.334383 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.334450 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.334473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.334504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.334527 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.343626 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/1.log" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.347115 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.348139 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.369288 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.391916 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.402870 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.416081 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.427084 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.437107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.437170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.437180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.437195 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.437205 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.438787 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.448688 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.456913 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.467647 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.479444 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.492703 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.525298 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.540769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.541114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.541231 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.541342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.541438 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.548648 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.582578 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.597924 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.614352 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.627227 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:11Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.644225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.644269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.644279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.644305 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.644317 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.746657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.746695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.746704 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.746720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.746730 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.848856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.848889 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.848906 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.848925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.848961 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.952336 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.952390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.952402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.952419 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:11 crc kubenswrapper[4926]: I1125 09:49:11.952430 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:11Z","lastTransitionTime":"2025-11-25T09:49:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.011999 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.012047 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.012130 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:12 crc kubenswrapper[4926]: E1125 09:49:12.012200 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:12 crc kubenswrapper[4926]: E1125 09:49:12.012320 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:12 crc kubenswrapper[4926]: E1125 09:49:12.012441 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.012730 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:12 crc kubenswrapper[4926]: E1125 09:49:12.012808 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.054793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.054837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.054847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.054863 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.054876 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.158515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.158576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.158595 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.158623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.158644 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.261418 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.261458 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.261469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.261490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.261502 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.354396 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/2.log" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.355628 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/1.log" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.359229 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f" exitCode=1 Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.359272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.359318 4926 scope.go:117] "RemoveContainer" containerID="3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.360282 4926 scope.go:117] "RemoveContainer" containerID="5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f" Nov 25 09:49:12 crc kubenswrapper[4926]: E1125 09:49:12.360499 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.365642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.365693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.365706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.365728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.365742 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.385103 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.403120 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.416250 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.436910 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.452123 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.468883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.468972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.468990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.469017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.469032 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.475562 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.508755 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3c5776ac9b16611e3544b2b0de99fea0ba792bd7cb8af76ca2d8d41f6ad76671\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:48:56Z\\\",\\\"message\\\":\\\":\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237900 6376 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-network-console/networking-console-plugin]} name:Service_openshift-network-console/networking-console-plugin_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.4.246:9443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {ab0b1d51-5ec6-479b-8881-93dfa8d30337}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1125 09:48:56.237971 6376 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238004 6376 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1125 09:48:56.238023 6376 factory.go:656] Stopping watch factory\\\\nI1125 09:48:56.238035 6376 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1125 09:48:56.238059 6376 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1125 09:48:56.238102 6376 ovnkube.go:599] Stopped ovnkube\\\\nI1125 09:48:56.238130 6376 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1125 09:48:56.238190 6376 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:55Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.524671 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.539723 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.551395 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.566155 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.571806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.571844 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.571855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.571875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.571887 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.579312 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.593713 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.606131 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.616925 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.630917 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.645417 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.674358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.674420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.674437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.674459 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.674471 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.777135 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.777172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.777181 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.777199 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.777210 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.880339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.880410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.880434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.880466 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.880489 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.983631 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.983669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.983676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.983692 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:12 crc kubenswrapper[4926]: I1125 09:49:12.983705 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:12Z","lastTransitionTime":"2025-11-25T09:49:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.087016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.087350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.087432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.087511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.087584 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.190080 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.190136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.190149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.190171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.190185 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.293044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.293678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.293789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.293890 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.293989 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.365998 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/2.log" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.371603 4926 scope.go:117] "RemoveContainer" containerID="5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f" Nov 25 09:49:13 crc kubenswrapper[4926]: E1125 09:49:13.371912 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.386303 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.396799 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.397096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.397168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.397238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.397305 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.400087 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.412009 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.428768 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.445209 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.467717 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.480150 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.497103 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.499712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.499747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.499756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.499771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.499781 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.518855 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.533587 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.551502 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.577420 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.595609 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.602376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.602432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.602443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.602464 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.602480 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.611383 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.623387 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.649074 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.663895 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:13Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.705198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.705270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.705289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.705317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.705335 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.808691 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.808741 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.808755 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.808778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.808795 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.911520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.911554 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.911582 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.911600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:13 crc kubenswrapper[4926]: I1125 09:49:13.911611 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:13Z","lastTransitionTime":"2025-11-25T09:49:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.011815 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.011965 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:14 crc kubenswrapper[4926]: E1125 09:49:14.012004 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.012122 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.012160 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:14 crc kubenswrapper[4926]: E1125 09:49:14.012214 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:14 crc kubenswrapper[4926]: E1125 09:49:14.012116 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:14 crc kubenswrapper[4926]: E1125 09:49:14.012298 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.013760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.013794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.013808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.013825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.013835 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.026526 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.038504 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.050987 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.061157 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.076024 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.097275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.110466 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.117297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.117327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.117335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.117352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.117365 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.123611 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.135684 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.150701 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.168920 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.181778 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.196322 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.209664 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.220459 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.220515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.220528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.220549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.220560 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.223280 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.239249 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.252278 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:14Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.254842 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:14 crc kubenswrapper[4926]: E1125 09:49:14.255085 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:14 crc kubenswrapper[4926]: E1125 09:49:14.255197 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:49:30.255176132 +0000 UTC m=+66.939132559 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.322778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.322820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.322829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.322846 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.322855 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.425407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.425789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.425916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.426047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.426114 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.529589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.529978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.530121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.530248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.530371 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.633548 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.633591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.633606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.633630 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.633643 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.736642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.737155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.737274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.737411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.737510 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.840599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.840665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.840686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.840714 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.840740 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.944393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.944474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.944499 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.944535 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:14 crc kubenswrapper[4926]: I1125 09:49:14.944559 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:14Z","lastTransitionTime":"2025-11-25T09:49:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.048067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.048118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.048128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.048146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.048157 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.178458 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.178517 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.178530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.178554 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.178569 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.281439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.281495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.281508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.281534 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.281547 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.384130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.384232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.384245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.384460 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.384473 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.487085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.487144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.487157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.487377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.487392 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.590675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.590747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.590765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.590976 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.591008 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.694249 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.694301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.694310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.694327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.694338 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.797513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.797563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.797574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.797597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.797611 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.883817 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.883964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.884010 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.884040 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884206 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884210 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:49:47.884168435 +0000 UTC m=+84.568124902 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884203 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884355 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:47.884328848 +0000 UTC m=+84.568285475 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884207 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884226 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884536 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:47.884508082 +0000 UTC m=+84.568464649 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884549 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.884615 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:47.884599083 +0000 UTC m=+84.568555510 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.900730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.900784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.900794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.900812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.900824 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:15Z","lastTransitionTime":"2025-11-25T09:49:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:15 crc kubenswrapper[4926]: I1125 09:49:15.985603 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.985806 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.985836 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.985851 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:15 crc kubenswrapper[4926]: E1125 09:49:15.985921 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:49:47.985902877 +0000 UTC m=+84.669859304 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.004143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.004217 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.004240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.004268 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.004286 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.011680 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.011777 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.011842 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:16 crc kubenswrapper[4926]: E1125 09:49:16.011905 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.012999 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:16 crc kubenswrapper[4926]: E1125 09:49:16.013151 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:16 crc kubenswrapper[4926]: E1125 09:49:16.012881 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:16 crc kubenswrapper[4926]: E1125 09:49:16.013342 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.107250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.107310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.107323 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.107343 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.107358 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.210284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.210342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.210351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.210369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.210380 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.313008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.313048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.313059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.313076 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.313089 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.415581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.415665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.415676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.415698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.415709 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.517991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.518034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.518048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.518067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.518078 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.620831 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.620870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.620878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.620894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.620904 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.723645 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.723683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.723693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.723710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.723720 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.826564 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.826612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.826622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.826641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.826654 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.929134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.929184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.929195 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.929216 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:16 crc kubenswrapper[4926]: I1125 09:49:16.929229 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:16Z","lastTransitionTime":"2025-11-25T09:49:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.032818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.032994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.033017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.033043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.033114 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.135523 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.135587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.135610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.135644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.135670 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.237793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.237857 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.237873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.237898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.237917 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.340675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.340711 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.340720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.340735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.340745 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.444086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.444156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.444172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.444198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.444215 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.546472 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.546528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.546539 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.546558 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.546572 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.650109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.650172 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.650189 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.650214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.650233 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.753707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.753747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.753757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.753778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.753790 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.856438 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.856509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.856520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.856544 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.856557 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.959475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.959532 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.959545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.959566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:17 crc kubenswrapper[4926]: I1125 09:49:17.959587 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:17Z","lastTransitionTime":"2025-11-25T09:49:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.011507 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.011501 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:18 crc kubenswrapper[4926]: E1125 09:49:18.011720 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.011538 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.011502 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:18 crc kubenswrapper[4926]: E1125 09:49:18.011972 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:18 crc kubenswrapper[4926]: E1125 09:49:18.012059 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:18 crc kubenswrapper[4926]: E1125 09:49:18.012260 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.062759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.062797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.062808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.062825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.062836 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.166355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.166413 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.166430 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.166461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.166477 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.269085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.269142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.269155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.269173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.269185 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.371572 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.371616 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.371625 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.371640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.371652 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.474461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.474499 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.474509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.474525 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.474536 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.577093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.577137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.577148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.577177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.577192 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.670689 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.680915 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.680982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.680998 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.681022 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.681040 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.683743 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.684265 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.703502 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.716543 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.727059 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.740591 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.760881 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.771751 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.781266 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.783970 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.784008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.784019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.784036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.784046 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.795737 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.809511 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.823696 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.834199 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.847859 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.861821 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.876325 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.886741 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.886800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.886811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.886831 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.886844 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.889929 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.904385 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:18Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.989026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.989096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.989105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.989123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:18 crc kubenswrapper[4926]: I1125 09:49:18.989139 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:18Z","lastTransitionTime":"2025-11-25T09:49:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.092683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.092758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.092771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.092794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.092808 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.195862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.195918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.195962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.195989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.196007 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.300002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.300049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.300059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.300077 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.300088 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.402577 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.402624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.402634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.402651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.402665 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.505329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.505389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.505400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.505423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.505445 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.608322 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.608363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.608374 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.608390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.608402 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.711465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.711503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.711522 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.711549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.711564 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.813691 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.813746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.813762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.813783 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.813800 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.917339 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.917415 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.917426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.917444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:19 crc kubenswrapper[4926]: I1125 09:49:19.917455 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:19Z","lastTransitionTime":"2025-11-25T09:49:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.012045 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.012111 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.012275 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.012280 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.012346 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.012372 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.012419 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.012470 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.019762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.019793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.019801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.019812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.019824 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.122840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.122881 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.122897 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.122915 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.122926 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.225264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.225318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.225330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.225353 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.225367 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.328927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.329052 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.329079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.329134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.329165 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.432067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.432141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.432161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.432188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.432205 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.535137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.535182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.535192 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.535212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.535227 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.591796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.591828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.591836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.591854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.591865 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.605115 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:20Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.611318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.611358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.611368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.611385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.611397 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.625838 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:20Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.629205 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.629252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.629262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.629279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.629319 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.641093 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:20Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.644888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.644915 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.644927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.644956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.644967 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.657156 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:20Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.660657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.660686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.660698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.660716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.660727 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.672283 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:20Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:20 crc kubenswrapper[4926]: E1125 09:49:20.672404 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.673922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.673969 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.673978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.673991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.674000 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.776597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.776634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.776644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.776662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.776673 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.880825 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.880892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.880915 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.880977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.881004 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.983653 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.983731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.983745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.983768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:20 crc kubenswrapper[4926]: I1125 09:49:20.983784 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:20Z","lastTransitionTime":"2025-11-25T09:49:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.087069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.087120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.087132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.087154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.087164 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.189738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.189779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.189788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.189806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.189817 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.293198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.293278 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.293302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.293335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.293359 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.396891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.397034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.397063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.397098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.397121 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.500216 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.500298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.500315 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.500337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.500349 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.603378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.603435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.603453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.603474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.603486 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.706479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.706527 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.706539 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.706559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.706573 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.808765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.808801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.808809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.808828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.808840 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.910816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.910860 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.910879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.910899 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:21 crc kubenswrapper[4926]: I1125 09:49:21.910911 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:21Z","lastTransitionTime":"2025-11-25T09:49:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.011557 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:22 crc kubenswrapper[4926]: E1125 09:49:22.011745 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.011557 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.011562 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.011759 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:22 crc kubenswrapper[4926]: E1125 09:49:22.011917 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:22 crc kubenswrapper[4926]: E1125 09:49:22.012012 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:22 crc kubenswrapper[4926]: E1125 09:49:22.012102 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.013786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.013840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.013917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.014010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.014054 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.117331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.117390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.117399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.117417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.117427 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.220133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.220195 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.220216 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.220237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.220249 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.323707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.323771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.323788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.323814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.323833 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.426230 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.426300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.426314 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.426334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.426347 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.530100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.530142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.530151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.530169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.530181 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.632608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.632653 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.632671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.632693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.632706 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.736245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.736302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.736312 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.736331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.736340 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.838695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.838735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.838747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.838768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.838780 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.941045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.941091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.941103 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.941124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:22 crc kubenswrapper[4926]: I1125 09:49:22.941137 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:22Z","lastTransitionTime":"2025-11-25T09:49:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.044572 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.044633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.044647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.044670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.044684 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.147817 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.147870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.147888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.147914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.147933 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.250829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.250877 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.250886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.250905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.250916 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.354431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.354504 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.354522 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.354556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.354589 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.457117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.457176 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.457189 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.457212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.457225 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.559882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.559922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.559981 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.560001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.560013 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.663984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.664054 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.664069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.664097 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.664118 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.767344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.767399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.767412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.767434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.767449 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.870604 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.870653 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.870665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.870683 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.870693 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.973147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.973230 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.973242 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.973262 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:23 crc kubenswrapper[4926]: I1125 09:49:23.973276 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:23Z","lastTransitionTime":"2025-11-25T09:49:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.011330 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:24 crc kubenswrapper[4926]: E1125 09:49:24.012025 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.013470 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.013490 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:24 crc kubenswrapper[4926]: E1125 09:49:24.013583 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.013609 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:24 crc kubenswrapper[4926]: E1125 09:49:24.013668 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:24 crc kubenswrapper[4926]: E1125 09:49:24.013883 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.027402 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.039919 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.053309 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.065016 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.076483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.076528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.076567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.076587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.076599 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.078660 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.093138 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.108197 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.120512 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.131129 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.150023 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.174309 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.179019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.179064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.179092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.179164 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.179189 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.187291 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.200555 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.214358 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.228888 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.241245 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.251852 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.265269 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:24Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.282522 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.282575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.282594 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.282617 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.282628 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.385086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.385147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.385161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.385192 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.385206 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.487845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.487895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.487905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.487921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.487933 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.590997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.591041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.591050 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.591065 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.591076 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.694285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.694342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.694355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.694377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.694392 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.797820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.797871 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.797882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.797901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.797912 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.900500 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.900567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.900604 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.900621 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:24 crc kubenswrapper[4926]: I1125 09:49:24.900632 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:24Z","lastTransitionTime":"2025-11-25T09:49:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.003600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.003647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.003658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.003677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.003689 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.106492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.106615 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.106677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.106710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.106774 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.209515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.209557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.209569 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.209587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.209598 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.312547 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.312590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.312599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.312616 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.312653 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.415067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.415111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.415119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.415140 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.415157 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.518505 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.518544 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.518577 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.518595 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.518605 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.620959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.621003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.621014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.621035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.621047 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.723575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.723636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.723648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.723669 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.723683 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.826879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.826952 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.826965 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.826989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.827003 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.930783 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.930855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.930868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.930888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:25 crc kubenswrapper[4926]: I1125 09:49:25.930904 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:25Z","lastTransitionTime":"2025-11-25T09:49:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.012449 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.012484 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.012536 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:26 crc kubenswrapper[4926]: E1125 09:49:26.012615 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:26 crc kubenswrapper[4926]: E1125 09:49:26.012770 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.012802 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:26 crc kubenswrapper[4926]: E1125 09:49:26.013000 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:26 crc kubenswrapper[4926]: E1125 09:49:26.013091 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.033804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.033842 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.033851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.033867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.033878 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.136763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.136811 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.136823 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.136840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.136852 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.239247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.239309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.239324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.239347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.239360 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.342043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.342147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.342167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.342191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.342204 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.445117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.445176 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.445188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.445206 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.445218 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.548452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.548501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.548514 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.548531 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.548547 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.651485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.651542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.651559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.651580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.651598 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.754607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.754650 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.754659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.754679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.754690 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.858143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.858184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.858194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.858212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.858223 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.961194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.961250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.961264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.961285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:26 crc kubenswrapper[4926]: I1125 09:49:26.961298 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:26Z","lastTransitionTime":"2025-11-25T09:49:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.012906 4926 scope.go:117] "RemoveContainer" containerID="5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f" Nov 25 09:49:27 crc kubenswrapper[4926]: E1125 09:49:27.013185 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.064226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.064272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.064281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.064296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.064308 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.167241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.167280 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.167290 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.167309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.167319 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.269986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.270044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.270055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.270075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.270087 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.375386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.375420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.375430 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.375446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.375461 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.478175 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.478250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.478276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.478389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.478450 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.580875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.580991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.581011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.581038 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.581059 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.683753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.683808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.683827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.683853 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.683870 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.786612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.786661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.786675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.786696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.786708 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.890030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.890089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.890101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.890119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.890133 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.992995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.993036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.993047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.993064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:27 crc kubenswrapper[4926]: I1125 09:49:27.993074 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:27Z","lastTransitionTime":"2025-11-25T09:49:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.011359 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.011394 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.011497 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.011540 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:28 crc kubenswrapper[4926]: E1125 09:49:28.011525 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:28 crc kubenswrapper[4926]: E1125 09:49:28.011706 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:28 crc kubenswrapper[4926]: E1125 09:49:28.011781 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:28 crc kubenswrapper[4926]: E1125 09:49:28.011863 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.095705 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.095747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.095756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.095775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.095786 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.198521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.198575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.198587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.198606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.198619 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.301482 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.301544 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.301556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.301577 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.301591 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.404886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.404957 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.404971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.404994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.405008 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.508010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.508166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.508180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.508209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.508241 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.611036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.611079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.611096 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.611117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.611130 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.714173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.714228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.714244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.714267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.714279 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.817664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.817722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.817732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.817756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.817774 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.920834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.920866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.920874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.920891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:28 crc kubenswrapper[4926]: I1125 09:49:28.920902 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:28Z","lastTransitionTime":"2025-11-25T09:49:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.023810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.023870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.023884 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.023908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.023928 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.126197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.126250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.126260 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.126279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.126291 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.229948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.230000 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.230012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.230032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.230045 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.333378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.333439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.333454 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.333474 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.333486 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.435919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.435974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.435984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.436002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.436013 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.538590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.538639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.538660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.538684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.538696 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.641021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.641067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.641076 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.641094 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.641108 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.743988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.744032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.744045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.744067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.744082 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.846116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.846160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.846173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.846193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.846206 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.948917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.948980 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.948991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.949009 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:29 crc kubenswrapper[4926]: I1125 09:49:29.949020 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:29Z","lastTransitionTime":"2025-11-25T09:49:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.012323 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.012365 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.012404 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.012512 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.012408 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.012591 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.012631 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.012685 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.052076 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.052118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.052127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.052144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.052154 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.155597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.155634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.155644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.155660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.155671 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.257920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.257977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.257989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.258005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.258016 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.355302 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.355624 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.355771 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:50:02.355738266 +0000 UTC m=+99.039694883 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.361160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.361202 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.361210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.361227 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.361241 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.464667 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.464732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.464745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.464772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.464789 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.567877 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.567925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.567961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.567981 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.567994 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.671171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.671238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.671255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.671316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.671338 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.774159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.774269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.774298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.774321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.774333 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.791222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.791276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.791286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.791306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.791318 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.805029 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.809633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.809688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.809703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.809726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.809738 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.825049 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.828899 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.828989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.829003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.829025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.829041 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.842419 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.846538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.846590 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.846608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.846628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.846641 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.857659 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.861851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.861874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.861884 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.861901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.861912 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.875991 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:30Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:30 crc kubenswrapper[4926]: E1125 09:49:30.876145 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.877731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.877827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.878138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.878225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.878415 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.982109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.982154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.982165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.982186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:30 crc kubenswrapper[4926]: I1125 09:49:30.982197 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:30Z","lastTransitionTime":"2025-11-25T09:49:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.084665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.084711 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.084721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.084738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.084750 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.187286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.187334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.187344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.187362 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.187373 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.290553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.290610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.290622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.290685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.290698 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.393473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.393515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.393530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.393546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.393567 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.497363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.497414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.497427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.497446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.497459 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.601003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.601074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.601089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.601109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.601122 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.704032 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.704084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.704102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.704124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.704138 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.806871 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.806912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.806924 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.806959 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.806973 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.909727 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.909784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.909796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.909815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:31 crc kubenswrapper[4926]: I1125 09:49:31.909828 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:31Z","lastTransitionTime":"2025-11-25T09:49:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.011651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.011842 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.011881 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.011924 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:32 crc kubenswrapper[4926]: E1125 09:49:32.012043 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:32 crc kubenswrapper[4926]: E1125 09:49:32.012164 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:32 crc kubenswrapper[4926]: E1125 09:49:32.012317 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:32 crc kubenswrapper[4926]: E1125 09:49:32.012369 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.014307 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.014386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.014406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.014431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.014449 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.117058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.117111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.117124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.117144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.117156 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.220079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.220128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.220138 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.220159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.220171 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.323115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.323164 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.323174 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.323193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.323202 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.425778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.425837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.425850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.425869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.425881 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.432496 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/0.log" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.432542 4926 generic.go:334] "Generic (PLEG): container finished" podID="cab27fcb-cde4-4879-8fad-79951474d7b3" containerID="53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff" exitCode=1 Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.432575 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerDied","Data":"53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.432993 4926 scope.go:117] "RemoveContainer" containerID="53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.448150 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.463490 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.481095 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.498100 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.510366 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.528892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.528972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.528982 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.529007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.529023 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.532923 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.546258 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.561131 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.574120 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.586926 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.602347 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.623887 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.631055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.631109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.631125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.631147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.631160 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.635495 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.648801 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.661867 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.675279 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.689235 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.706129 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:32Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.734016 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.734057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.734067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.734086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.734099 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.837093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.837134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.837143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.837160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.837170 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.940132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.940184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.940198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.940221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:32 crc kubenswrapper[4926]: I1125 09:49:32.940235 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:32Z","lastTransitionTime":"2025-11-25T09:49:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.043732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.043785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.043798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.043819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.043832 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.146541 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.146580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.146588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.146603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.146612 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.249437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.249494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.249511 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.249533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.249551 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.352699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.352750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.352768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.352794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.352820 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.440267 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/0.log" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.440358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerStarted","Data":"1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.455058 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.456356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.456402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.456418 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.456439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.456454 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.467313 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.488412 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.500495 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.517590 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.533498 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.555030 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.559151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.559206 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.559221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.559243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.559256 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.568894 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.582094 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.592766 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.606901 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.622590 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.643311 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.657244 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.661264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.661308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.661320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.661341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.661351 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.667833 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.684116 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.698023 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.709623 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:33Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.763439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.763495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.763510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.763537 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.763550 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.866404 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.866468 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.866483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.866510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.866522 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.973083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.973132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.973142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.973165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:33 crc kubenswrapper[4926]: I1125 09:49:33.973176 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:33Z","lastTransitionTime":"2025-11-25T09:49:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.012272 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:34 crc kubenswrapper[4926]: E1125 09:49:34.012739 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.013033 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:34 crc kubenswrapper[4926]: E1125 09:49:34.013235 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.013262 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:34 crc kubenswrapper[4926]: E1125 09:49:34.013432 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.013315 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:34 crc kubenswrapper[4926]: E1125 09:49:34.013621 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.031471 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.044851 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.059758 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.075925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.076222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.076283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.076475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.076554 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.084162 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.102907 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.117093 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.135132 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.157609 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.172883 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.178849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.178892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.178904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.178923 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.178955 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.185332 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.196871 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.213554 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.228374 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.242014 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.254577 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.267289 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.282107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.282151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.282162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.282061 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.282180 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.282293 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.296823 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:34Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.385685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.385735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.385745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.385764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.385777 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.489159 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.489293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.489324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.489346 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.489383 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.596850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.596918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.596928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.596962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.596973 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.699536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.699602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.699614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.699634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.699646 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.802798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.802875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.802893 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.802930 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.802988 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.905587 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.905628 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.905639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.905658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:34 crc kubenswrapper[4926]: I1125 09:49:34.905671 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:34Z","lastTransitionTime":"2025-11-25T09:49:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.007838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.007880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.007888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.007917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.007928 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.110313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.110358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.110370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.110390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.110403 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.213244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.213293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.213338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.213368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.213408 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.316704 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.316752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.316762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.316781 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.316792 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.419677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.419726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.419739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.419758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.419770 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.522695 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.522746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.522759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.522782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.522798 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.625053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.625099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.625112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.625130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.625141 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.727431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.727477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.727486 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.727503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.727516 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.830250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.830301 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.830313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.830332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.830347 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.933726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.933792 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.933805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.933824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:35 crc kubenswrapper[4926]: I1125 09:49:35.933837 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:35Z","lastTransitionTime":"2025-11-25T09:49:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.011646 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.011708 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.011744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.011654 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:36 crc kubenswrapper[4926]: E1125 09:49:36.011833 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:36 crc kubenswrapper[4926]: E1125 09:49:36.011955 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:36 crc kubenswrapper[4926]: E1125 09:49:36.012024 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:36 crc kubenswrapper[4926]: E1125 09:49:36.012068 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.036656 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.036701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.036713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.036734 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.036748 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.139418 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.139462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.139471 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.139488 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.139501 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.241978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.242027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.242038 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.242057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.242068 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.344713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.344767 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.344782 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.344803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.344816 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.447747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.447814 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.447826 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.447843 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.447858 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.551132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.551216 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.551228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.551252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.551265 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.653772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.653829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.653846 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.653870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.653883 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.757574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.757618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.757626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.757644 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.757656 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.860280 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.860326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.860340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.860386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.860405 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.963992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.964060 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.964077 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.964100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:36 crc kubenswrapper[4926]: I1125 09:49:36.964115 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:36Z","lastTransitionTime":"2025-11-25T09:49:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.071029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.071202 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.071264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.071300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.071465 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.175834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.175887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.175900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.175921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.175967 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.278568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.278619 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.278631 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.278649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.278662 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.381783 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.381840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.381853 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.381874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.381889 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.484749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.484810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.484827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.484861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.484880 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.588044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.588107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.588122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.588144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.588158 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.690708 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.690759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.690775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.690802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.690818 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.794843 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.794901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.794917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.794968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.794987 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.898167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.898209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.898225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.898243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:37 crc kubenswrapper[4926]: I1125 09:49:37.898254 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:37Z","lastTransitionTime":"2025-11-25T09:49:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.001294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.001371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.001386 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.001406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.001441 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.011836 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.011865 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.011835 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:38 crc kubenswrapper[4926]: E1125 09:49:38.011981 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:38 crc kubenswrapper[4926]: E1125 09:49:38.012030 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.012063 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:38 crc kubenswrapper[4926]: E1125 09:49:38.012099 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:38 crc kubenswrapper[4926]: E1125 09:49:38.012220 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.104706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.104758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.104768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.104785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.104796 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.207521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.207578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.207596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.207626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.207652 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.310244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.310295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.310310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.310327 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.310340 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.413557 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.413633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.413658 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.413692 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.413719 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.517920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.517992 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.518007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.518030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.518044 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.620882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.620978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.621002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.621036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.621060 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.724143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.724197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.724208 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.724228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.724241 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.827712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.827766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.827776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.827794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.827805 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.930769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.930969 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.930987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.931012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:38 crc kubenswrapper[4926]: I1125 09:49:38.931027 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:38Z","lastTransitionTime":"2025-11-25T09:49:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.033772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.033859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.033891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.033918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.034011 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.136476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.136527 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.136540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.136555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.136566 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.239431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.239484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.239494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.239512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.239527 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.343286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.343361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.343406 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.343441 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.343460 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.447170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.447250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.447275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.447303 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.447321 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.549885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.549929 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.549987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.550013 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.550027 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.653193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.653257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.653274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.653299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.653317 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.758323 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.758380 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.758394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.758417 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.758432 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.860672 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.860726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.860739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.860758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.860771 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.963031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.963092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.963104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.963122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:39 crc kubenswrapper[4926]: I1125 09:49:39.963132 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:39Z","lastTransitionTime":"2025-11-25T09:49:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.011442 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:40 crc kubenswrapper[4926]: E1125 09:49:40.011630 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.011883 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:40 crc kubenswrapper[4926]: E1125 09:49:40.011975 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.012178 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.012192 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:40 crc kubenswrapper[4926]: E1125 09:49:40.012354 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:40 crc kubenswrapper[4926]: E1125 09:49:40.012570 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.066062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.066113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.066132 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.066153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.066165 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.169061 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.169114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.169124 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.169148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.169160 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.294562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.294615 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.294627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.294656 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.294670 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.397804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.397849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.397861 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.397880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.397891 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.500987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.501034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.501044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.501062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.501073 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.604001 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.604070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.604093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.604122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.604138 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.707304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.707373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.707393 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.707424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.707443 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.809934 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.810005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.810019 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.810041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.810055 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.912822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.912901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.912921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.913014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:40 crc kubenswrapper[4926]: I1125 09:49:40.913034 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:40Z","lastTransitionTime":"2025-11-25T09:49:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.013367 4926 scope.go:117] "RemoveContainer" containerID="5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.015987 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.016049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.016063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.016087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.016102 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.053693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.053737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.053749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.053769 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.053783 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: E1125 09:49:41.066362 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.071047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.071113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.071126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.071142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.071154 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: E1125 09:49:41.086656 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.090856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.090896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.090909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.090948 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.090963 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: E1125 09:49:41.105747 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.110292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.110325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.110337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.110363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.110377 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: E1125 09:49:41.123484 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.127318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.127359 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.127372 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.127392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.127404 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: E1125 09:49:41.150058 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: E1125 09:49:41.150270 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.152173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.152212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.152229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.152299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.152318 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.255045 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.255087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.255098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.255115 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.255126 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.358005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.358068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.358086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.358113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.358135 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.461618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.461660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.461670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.461685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.461696 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.469048 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/2.log" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.471741 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.472184 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.487316 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.507880 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.522706 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.534709 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.546959 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.564610 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.564662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.564676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.564700 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.564715 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.565602 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.580392 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.596424 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.611220 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.623243 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.636247 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.649709 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.661886 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.667203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.667236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.667247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.667263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.667272 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.675399 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.687322 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.708888 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.721560 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.735645 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.769720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.769768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.769779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.769797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.769806 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.872904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.872976 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.872989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.873010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.873024 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.976344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.976391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.976431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.976454 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:41 crc kubenswrapper[4926]: I1125 09:49:41.976465 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:41Z","lastTransitionTime":"2025-11-25T09:49:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.012281 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.012325 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.012437 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:42 crc kubenswrapper[4926]: E1125 09:49:42.012456 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.012511 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:42 crc kubenswrapper[4926]: E1125 09:49:42.012665 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:42 crc kubenswrapper[4926]: E1125 09:49:42.012693 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:42 crc kubenswrapper[4926]: E1125 09:49:42.012848 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.079316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.079362 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.079375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.079396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.079411 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.183727 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.183802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.183820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.183848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.183867 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.286784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.286849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.286877 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.286913 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.286968 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.390802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.390859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.390873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.390896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.390911 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.478473 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/3.log" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.479324 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/2.log" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.483506 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" exitCode=1 Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.483569 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.483630 4926 scope.go:117] "RemoveContainer" containerID="5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.485731 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:49:42 crc kubenswrapper[4926]: E1125 09:49:42.485971 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.495056 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.495107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.495126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.495152 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.495169 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.502856 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.524379 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.538637 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.557037 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.571182 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.585772 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.597495 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.601177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.601218 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.601232 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.601257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.601271 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.611503 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.633412 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.646048 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.655236 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.675119 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.704522 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5a7180cbf81ad69df26ddb565e97110fed90520a35271a8737adce110d29754f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:12Z\\\",\\\"message\\\":\\\"646444c4-trplf\\\\nI1125 09:49:12.036288 6592 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1125 09:49:12.036297 6592 obj_retry.go:303] Retry object setup: *v1.Pod openshift-kube-controller-manager/kube-controller-manager-crc\\\\nI1125 09:49:12.035866 6592 obj_retry.go:365] Adding new object: *v1.Pod openshift-dns/node-resolver-58wwc\\\\nI1125 09:49:12.036337 6592 base_network_controller_pods.go:916] Annotation values: ip=[10.217.0.3/23] ; mac=0a:58:0a:d9:00:03 ; gw=[10.217.0.1]\\\\nI1125 09:49:12.036362 6592 ovn.go:134] Ensuring zone local for Pod openshift-dns/node-resolver-58wwc in node crc\\\\nI1125 09:49:12.036377 6592 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-58wwc after 0 failed attempt(s)\\\\nI1125 09:49:12.036387 6592 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-58wwc\\\\nF1125 09:49:12.036264 6592 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to star\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"rc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z]\\\\nI1125 09:49:41.821139 6955 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.705275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.705316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.705326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.705344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.705355 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.721989 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.737654 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.752228 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.767892 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.781978 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:42Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.808341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.808378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.808390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.808411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.808423 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.911802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.911853 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.911869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.911891 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:42 crc kubenswrapper[4926]: I1125 09:49:42.911908 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:42Z","lastTransitionTime":"2025-11-25T09:49:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.014915 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.014984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.014999 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.015021 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.015038 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.029300 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.118988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.119077 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.119102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.119136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.119163 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.222088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.222142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.222154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.222181 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.222196 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.325928 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.326023 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.326040 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.326062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.326079 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.429079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.429414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.429424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.429442 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.429471 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.489468 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/3.log" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.496419 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:49:43 crc kubenswrapper[4926]: E1125 09:49:43.496631 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.515427 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.528709 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.533008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.533048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.533072 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.533095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.533110 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.541769 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.563238 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.580376 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.595814 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.612815 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.632597 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"rc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z]\\\\nI1125 09:49:41.821139 6955 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.637070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.637155 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.637170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.637553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.637594 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.648958 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.661745 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.672667 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.686222 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.698786 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.711917 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.725185 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.740681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.740742 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.740757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.740779 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.740790 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.741240 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.759655 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.771537 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"086ce9d7-88bf-4bab-a5b7-5083fefd3d14\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a386f401a9684c2ea8c4941f2f3bbb17051796ccfe35e00841bb3426bc531e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.785275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:43Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.843962 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.843997 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.844008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.844029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.844040 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.947107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.947170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.947187 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.947210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:43 crc kubenswrapper[4926]: I1125 09:49:43.947226 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:43Z","lastTransitionTime":"2025-11-25T09:49:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.011733 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.011733 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.011746 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:44 crc kubenswrapper[4926]: E1125 09:49:44.011899 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.012049 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:44 crc kubenswrapper[4926]: E1125 09:49:44.012119 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:44 crc kubenswrapper[4926]: E1125 09:49:44.012206 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:44 crc kubenswrapper[4926]: E1125 09:49:44.012331 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.026371 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.039130 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.050053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.050100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.050133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.050153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.050164 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.051890 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.063337 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"086ce9d7-88bf-4bab-a5b7-5083fefd3d14\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a386f401a9684c2ea8c4941f2f3bbb17051796ccfe35e00841bb3426bc531e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.078861 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.093322 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.107035 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.118352 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.137275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.148275 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.154569 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.154638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.154652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.154675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.154688 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.161391 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.174677 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.200334 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"rc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z]\\\\nI1125 09:49:41.821139 6955 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.215153 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.227277 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.239117 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.254892 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.256983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.257017 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.257026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.257059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.257077 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.270659 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.284366 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:44Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.361194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.361244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.361256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.361275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.361287 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.463926 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.463998 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.464010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.464049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.464064 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.567251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.567294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.567340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.567361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.567382 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.671036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.671087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.671108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.671125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.671135 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.773837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.773883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.773893 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.773912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.773924 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.876909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.876977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.876989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.877007 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.877020 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.980147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.980221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.980239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.980259 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:44 crc kubenswrapper[4926]: I1125 09:49:44.980290 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:44Z","lastTransitionTime":"2025-11-25T09:49:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.083146 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.083229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.083240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.083257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.083268 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.186421 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.186476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.186488 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.186508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.186520 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.289840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.289895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.289910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.289933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.289966 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.393129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.393181 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.393191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.393229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.393245 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.496636 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.496685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.496697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.496722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.496735 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.600168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.600237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.600252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.600275 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.600288 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.703089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.703144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.703157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.703177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.703189 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.806876 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.806977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.807004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.807041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.807067 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.909750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.909793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.909807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.909824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:45 crc kubenswrapper[4926]: I1125 09:49:45.909835 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:45Z","lastTransitionTime":"2025-11-25T09:49:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012259 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012325 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:46 crc kubenswrapper[4926]: E1125 09:49:46.012535 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012387 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012333 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012713 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.012729 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: E1125 09:49:46.012753 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:46 crc kubenswrapper[4926]: E1125 09:49:46.012806 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:46 crc kubenswrapper[4926]: E1125 09:49:46.012705 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.116851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.116923 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.116956 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.116977 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.116989 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.220293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.220363 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.220376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.220401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.220421 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.323990 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.324050 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.324068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.324089 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.324102 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.428657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.428722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.428735 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.428753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.428764 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.532317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.532371 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.532385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.532403 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.532416 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.635492 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.635551 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.635592 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.635611 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.635623 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.739716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.739774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.739791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.739817 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.739835 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.842796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.842870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.842885 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.842908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.842921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.946064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.946125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.946141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.946165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:46 crc kubenswrapper[4926]: I1125 09:49:46.946178 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:46Z","lastTransitionTime":"2025-11-25T09:49:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.049097 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.049156 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.049168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.049186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.049196 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.152726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.152774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.152790 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.152809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.152822 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.255710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.255744 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.255760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.255778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.255789 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.358839 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.358927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.358988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.359024 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.359048 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.461741 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.461809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.461820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.461836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.461846 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.564854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.564922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.564965 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.564995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.565009 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.667640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.667693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.667702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.667719 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.667732 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.770723 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.770765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.770775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.770793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.770804 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.873213 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.873263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.873279 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.873304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.873319 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.968328 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.968510 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.968578 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.968544102 +0000 UTC m=+148.652500569 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.968653 4926 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.968678 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.968739 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.968712737 +0000 UTC m=+148.652669214 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.968779 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.968829 4926 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.968919 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.968891412 +0000 UTC m=+148.652847879 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.969215 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.969255 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.969283 4926 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:47 crc kubenswrapper[4926]: E1125 09:49:47.969370 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.969345215 +0000 UTC m=+148.653301682 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.976591 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.976674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.976701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.976731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:47 crc kubenswrapper[4926]: I1125 09:49:47.976754 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:47Z","lastTransitionTime":"2025-11-25T09:49:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.011965 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.012044 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.012067 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.012185 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.012546 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.012663 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.013142 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.013354 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.070149 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.070447 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.070512 4926 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.070541 4926 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:48 crc kubenswrapper[4926]: E1125 09:49:48.070642 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.0706152 +0000 UTC m=+148.754571667 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.080883 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.080983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.081012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.081043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.082031 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.185638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.185692 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.185708 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.185733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.185756 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.288759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.289101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.289112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.289131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.289142 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.391752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.391805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.391816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.391832 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.391842 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.494751 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.494809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.494822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.494845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.494856 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.597453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.597545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.597556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.597573 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.597583 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.700599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.700862 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.700886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.700905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.700917 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.804512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.804546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.804556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.804574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.804587 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.908345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.908396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.908405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.908424 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:48 crc kubenswrapper[4926]: I1125 09:49:48.908436 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:48Z","lastTransitionTime":"2025-11-25T09:49:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.011798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.011889 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.011909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.011933 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.011996 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.113620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.113656 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.113666 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.113682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.113693 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.216543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.216585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.216597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.216614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.216630 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.320528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.320699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.320724 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.320759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.320781 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.423523 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.423564 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.423578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.423599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.423613 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.526469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.526533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.526543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.526560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.526570 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.629186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.629238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.629252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.629272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.629285 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.732978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.733071 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.733088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.733112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.733126 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.836824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.837173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.837278 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.837347 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.837409 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.940794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.940845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.940858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.940882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:49 crc kubenswrapper[4926]: I1125 09:49:49.940897 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:49Z","lastTransitionTime":"2025-11-25T09:49:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.011974 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.012006 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.011974 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.012109 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:50 crc kubenswrapper[4926]: E1125 09:49:50.012177 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:50 crc kubenswrapper[4926]: E1125 09:49:50.012272 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:50 crc kubenswrapper[4926]: E1125 09:49:50.012477 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:50 crc kubenswrapper[4926]: E1125 09:49:50.012539 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.043391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.043432 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.043444 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.043462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.043474 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.146566 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.146623 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.146641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.146662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.146674 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.249498 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.249865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.249930 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.250052 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.250112 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.352437 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.352482 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.352491 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.352508 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.352521 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.455209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.455252 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.455273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.455293 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.455305 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.558055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.558128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.558143 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.558167 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.558181 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.660838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.661216 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.661296 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.661390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.661459 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.765703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.765750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.765765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.765788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.765809 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.868677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.868738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.868750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.868773 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.868785 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.972267 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.972303 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.972311 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.972326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:50 crc kubenswrapper[4926]: I1125 09:49:50.972335 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:50Z","lastTransitionTime":"2025-11-25T09:49:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.076435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.076529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.076559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.076593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.076616 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.180137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.180200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.180209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.180228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.180245 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.283023 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.283073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.283087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.283109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.283123 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.385622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.385668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.385677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.385697 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.385715 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.468303 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.468355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.468370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.468388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.468401 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: E1125 09:49:51.498682 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.504325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.504377 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.504388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.504410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.504423 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: E1125 09:49:51.525295 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.531185 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.531241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.531253 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.531274 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.531289 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: E1125 09:49:51.545839 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.551212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.551261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.551273 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.551291 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.551301 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: E1125 09:49:51.563751 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.568035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.568084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.568094 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.568114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.568127 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: E1125 09:49:51.580296 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:51Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:51 crc kubenswrapper[4926]: E1125 09:49:51.580411 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.582141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.582188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.582202 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.582220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.582233 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.684670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.684731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.684745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.684770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.684784 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.787690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.787750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.787770 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.787793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.787841 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.891349 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.891388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.891397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.891412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.891422 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.994329 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.994382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.994397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.994414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:51 crc kubenswrapper[4926]: I1125 09:49:51.994424 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:51Z","lastTransitionTime":"2025-11-25T09:49:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.012270 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.012346 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:52 crc kubenswrapper[4926]: E1125 09:49:52.012456 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.012517 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.012591 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:52 crc kubenswrapper[4926]: E1125 09:49:52.012583 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:52 crc kubenswrapper[4926]: E1125 09:49:52.012812 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:52 crc kubenswrapper[4926]: E1125 09:49:52.012890 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.097523 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.097583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.097607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.097633 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.097654 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.200836 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.200888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.200902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.200923 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.200958 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.304546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.304606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.304626 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.304645 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.304659 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.407539 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.407589 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.407601 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.407624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.407641 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.511401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.511451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.511465 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.511513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.511530 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.613958 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.614048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.614069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.614094 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.614108 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.717407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.717473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.717485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.717506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.717522 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.820501 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.820565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.820593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.820621 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.820636 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.924489 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.924742 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.924824 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.924880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:52 crc kubenswrapper[4926]: I1125 09:49:52.924907 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:52Z","lastTransitionTime":"2025-11-25T09:49:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.028127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.028190 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.028200 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.028220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.028232 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.131350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.131382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.131390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.131409 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.131418 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.234095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.234131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.234145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.234162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.234174 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.337539 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.337606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.337653 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.337677 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.337692 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.441087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.441531 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.441607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.441717 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.441790 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.543766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.543808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.543819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.543838 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.543851 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.646518 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.646552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.646559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.646576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.646588 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.750098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.750148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.750171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.750203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.750220 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.852904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.852971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.852985 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.853006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.853021 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.955921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.955995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.956009 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.956034 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:53 crc kubenswrapper[4926]: I1125 09:49:53.956049 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:53Z","lastTransitionTime":"2025-11-25T09:49:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.013645 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.013729 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.013729 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:54 crc kubenswrapper[4926]: E1125 09:49:54.013810 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.013866 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:54 crc kubenswrapper[4926]: E1125 09:49:54.014067 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:54 crc kubenswrapper[4926]: E1125 09:49:54.014169 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:54 crc kubenswrapper[4926]: E1125 09:49:54.014615 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.029994 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.050073 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"rc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z]\\\\nI1125 09:49:41.821139 6955 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.058971 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.059292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.059423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.059521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.059615 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.062533 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.078316 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.093355 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.109483 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.120458 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.131237 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.142000 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"086ce9d7-88bf-4bab-a5b7-5083fefd3d14\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a386f401a9684c2ea8c4941f2f3bbb17051796ccfe35e00841bb3426bc531e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.155774 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.161786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.161822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.161832 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.161849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.161861 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.169530 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.184550 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.197860 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.212678 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.234239 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.248775 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.262251 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.264333 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.264423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.264446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.264495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.264514 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.275332 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.285044 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:54Z is after 2025-08-24T17:21:41Z" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.368160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.368534 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.368810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.368896 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.369001 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.473310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.473362 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.473375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.473397 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.473408 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.575870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.575912 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.575922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.575958 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.575970 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.678833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.679585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.679663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.679743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.679818 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.783175 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.783222 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.783233 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.783250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.783264 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.886435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.886515 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.886528 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.886553 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.886566 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.989355 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.989414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.989429 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.989452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:54 crc kubenswrapper[4926]: I1125 09:49:54.989464 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:54Z","lastTransitionTime":"2025-11-25T09:49:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.092506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.092568 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.092580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.092598 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.092609 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.195707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.195755 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.195765 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.195784 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.195796 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.299226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.299278 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.299289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.299308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.299322 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.401771 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.401818 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.401830 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.401851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.401864 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.505039 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.505097 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.505108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.505128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.505142 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.607681 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.607750 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.607760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.607780 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.607799 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.710204 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.710263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.710277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.710299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.710315 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.812991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.813046 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.813055 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.813075 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.813086 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.916536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.916575 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.916584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.916599 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:55 crc kubenswrapper[4926]: I1125 09:49:55.916608 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:55Z","lastTransitionTime":"2025-11-25T09:49:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.011783 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.011818 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.011794 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.011855 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:56 crc kubenswrapper[4926]: E1125 09:49:56.011999 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:56 crc kubenswrapper[4926]: E1125 09:49:56.012110 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:56 crc kubenswrapper[4926]: E1125 09:49:56.012128 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:56 crc kubenswrapper[4926]: E1125 09:49:56.012199 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.018451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.018494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.018510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.018530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.018543 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.120622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.120667 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.120680 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.120701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.120715 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.223758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.223808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.223817 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.223832 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.223842 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.326894 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.326996 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.327011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.327033 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.327048 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.430161 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.430213 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.430223 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.430240 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.430251 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.533266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.533772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.534084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.534112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.534128 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.637925 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.638322 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.638392 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.638467 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.638543 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.740999 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.741661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.741759 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.741859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.741951 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.844690 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.845030 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.845151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.845246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.845340 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.948388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.948426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.948435 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.948451 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:56 crc kubenswrapper[4926]: I1125 09:49:56.948460 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:56Z","lastTransitionTime":"2025-11-25T09:49:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.012228 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:49:57 crc kubenswrapper[4926]: E1125 09:49:57.012441 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.051901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.052348 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.052461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.052576 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.052667 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.155543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.155584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.155596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.155614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.155625 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.257960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.257995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.258008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.258026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.258037 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.361160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.361194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.361202 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.361217 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.361227 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.464198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.464236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.464245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.464261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.464271 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.567003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.567037 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.567048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.567066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.567081 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.670068 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.670125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.670137 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.670191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.670205 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.772733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.772777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.772786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.772808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.772826 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.875530 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.875919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.876074 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.876235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.876368 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.978863 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.978911 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.978922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.978960 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:57 crc kubenswrapper[4926]: I1125 09:49:57.978976 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:57Z","lastTransitionTime":"2025-11-25T09:49:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.012108 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.012138 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.012168 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.012227 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:49:58 crc kubenswrapper[4926]: E1125 09:49:58.012299 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:49:58 crc kubenswrapper[4926]: E1125 09:49:58.012400 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:49:58 crc kubenswrapper[4926]: E1125 09:49:58.012490 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:49:58 crc kubenswrapper[4926]: E1125 09:49:58.012581 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.081637 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.081674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.081687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.081707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.081719 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.183753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.183815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.183827 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.183845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.183856 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.286286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.286684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.286809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.286901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.287032 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.390203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.390236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.390246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.390260 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.390270 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.492284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.492330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.492344 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.492361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.492375 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.595128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.595484 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.595607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.595699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.595777 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.699127 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.699191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.699203 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.699228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.699243 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.802070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.802145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.802162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.802184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.802197 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.905057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.905476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.905567 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.905663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:58 crc kubenswrapper[4926]: I1125 09:49:58.905738 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:58Z","lastTransitionTime":"2025-11-25T09:49:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.008263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.008309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.008321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.008340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.008350 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.110787 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.110847 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.110858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.110877 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.110889 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.213008 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.213036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.213044 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.213059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.213068 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.315841 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.315882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.315893 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.315910 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.315921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.418641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.418974 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.419051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.419129 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.419201 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.522106 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.522503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.522609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.522710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.522789 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.625570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.625608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.625620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.625641 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.625654 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.728510 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.728560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.728574 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.728597 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.728612 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.831737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.831793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.831810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.831832 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.831844 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.935419 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.935533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.935552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.935730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:49:59 crc kubenswrapper[4926]: I1125 09:49:59.935815 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:49:59Z","lastTransitionTime":"2025-11-25T09:49:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.011472 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.011496 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.011520 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:00 crc kubenswrapper[4926]: E1125 09:50:00.011626 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:00 crc kubenswrapper[4926]: E1125 09:50:00.011717 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:00 crc kubenswrapper[4926]: E1125 09:50:00.011833 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.012486 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:00 crc kubenswrapper[4926]: E1125 09:50:00.012601 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.039375 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.039431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.039449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.039469 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.039483 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.142787 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.142859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.142869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.142888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.142901 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.246063 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.246113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.246128 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.246148 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.246174 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.348671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.349027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.349048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.349064 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.349103 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.452728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.452788 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.452798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.452816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.452827 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.555070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.555130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.555145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.555165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.555179 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.657602 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.657646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.657657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.657675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.657685 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.760005 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.760053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.760066 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.760088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.760107 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.863321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.863382 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.863402 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.863429 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.863447 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.965993 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.966027 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.966035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.966051 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:00 crc kubenswrapper[4926]: I1125 09:50:00.966062 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:00Z","lastTransitionTime":"2025-11-25T09:50:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.068648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.068703 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.068718 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.068741 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.068759 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.172048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.172099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.172133 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.172154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.172165 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.274565 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.274620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.274634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.274654 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.274665 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.377729 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.377766 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.377776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.377791 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.377800 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.481166 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.481230 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.481244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.481264 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.481277 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.583797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.583859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.583870 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.583890 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.583902 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.686364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.686403 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.686412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.686429 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.686440 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.789737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.789787 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.789795 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.789812 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.789823 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.830663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.830716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.830730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.830756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.830769 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: E1125 09:50:01.844492 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.848807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.848854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.848868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.848898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.848913 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: E1125 09:50:01.862542 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.867358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.867407 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.867422 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.867446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.867461 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: E1125 09:50:01.881138 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.885876 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.885917 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.885953 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.885978 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.885994 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: E1125 09:50:01.898456 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.903067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.903105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.903114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.903130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.903141 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:01 crc kubenswrapper[4926]: E1125 09:50:01.916384 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:01Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:01 crc kubenswrapper[4926]: E1125 09:50:01.916558 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.918585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.918634 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.918646 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.918664 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:01 crc kubenswrapper[4926]: I1125 09:50:01.918678 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:01Z","lastTransitionTime":"2025-11-25T09:50:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.011587 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.011676 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.011617 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.011593 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:02 crc kubenswrapper[4926]: E1125 09:50:02.011756 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:02 crc kubenswrapper[4926]: E1125 09:50:02.011853 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:02 crc kubenswrapper[4926]: E1125 09:50:02.011930 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:02 crc kubenswrapper[4926]: E1125 09:50:02.012031 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.020831 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.020887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.020901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.020921 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.020957 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.124101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.124157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.124168 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.124194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.124209 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.226496 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.226536 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.226545 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.226561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.226571 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.329774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.329904 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.329922 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.330287 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.330306 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.424183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:02 crc kubenswrapper[4926]: E1125 09:50:02.424342 4926 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:50:02 crc kubenswrapper[4926]: E1125 09:50:02.424456 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs podName:81f30c43-dacf-407f-adb3-bdbae866485f nodeName:}" failed. No retries permitted until 2025-11-25 09:51:06.424436258 +0000 UTC m=+163.108392685 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs") pod "network-metrics-daemon-xpn28" (UID: "81f30c43-dacf-407f-adb3-bdbae866485f") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.432874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.432909 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.432918 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.432961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.432974 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.536079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.536131 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.536150 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.536173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.536187 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.638932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.639000 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.639012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.639031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.639041 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.741627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.741689 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.741706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.741732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.741754 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.844446 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.844494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.844503 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.844521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.844535 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.948049 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.948091 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.948102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.948123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:02 crc kubenswrapper[4926]: I1125 09:50:02.948133 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:02Z","lastTransitionTime":"2025-11-25T09:50:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.051249 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.051304 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.051316 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.051340 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.051369 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.153772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.153859 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.153873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.153890 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.153900 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.257284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.257324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.257341 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.257357 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.257369 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.361261 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.361318 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.361332 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.361356 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.361372 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.463908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.464004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.464020 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.464039 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.464052 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.567902 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.567961 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.567991 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.568010 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.568021 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.670497 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.670556 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.670572 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.670595 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.670611 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.773053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.773122 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.773134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.773157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.773169 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.876475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.876512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.876540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.876569 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.876580 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.980642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.980699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.980710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.980727 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:03 crc kubenswrapper[4926]: I1125 09:50:03.980738 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:03Z","lastTransitionTime":"2025-11-25T09:50:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.012309 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.012378 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.012327 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.012324 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:04 crc kubenswrapper[4926]: E1125 09:50:04.013048 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:04 crc kubenswrapper[4926]: E1125 09:50:04.013220 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:04 crc kubenswrapper[4926]: E1125 09:50:04.013329 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:04 crc kubenswrapper[4926]: E1125 09:50:04.013405 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.028398 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-l78n4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cab27fcb-cde4-4879-8fad-79951474d7b3\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:32Z\\\",\\\"message\\\":\\\"2025-11-25T09:48:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987\\\\n2025-11-25T09:48:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_02407748-9801-4762-a8d2-d2e534ed3987 to /host/opt/cni/bin/\\\\n2025-11-25T09:48:47Z [verbose] multus-daemon started\\\\n2025-11-25T09:48:47Z [verbose] Readiness Indicator file check\\\\n2025-11-25T09:49:32Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:49:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-h6mn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-l78n4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.040863 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"086ce9d7-88bf-4bab-a5b7-5083fefd3d14\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a386f401a9684c2ea8c4941f2f3bbb17051796ccfe35e00841bb3426bc531e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2287e62b41340a808ecce6455351a0996473373b41e14c2cd9dac10863f4f3bf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.056989 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f5f3eeede30718ddc38633deca4312848c30787dd79fd66a79644f863ea0a899\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.071735 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.083753 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.083794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.083803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.083820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.083834 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.086035 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.096674 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-58wwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a918afbd-3489-49de-a796-b76a7d73748c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7819f44fbeed033b2cb68fc7e9c28ee07083ae73b6af4790ec3eda7ca455b0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hrmp6\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-58wwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.119383 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dbbdfb7-479c-43e4-a138-37df380a6399\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cbd5f6976b182ad998b5f48310ebe92a99cdd1d5e7150f9905a69ba4c4b67eaa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3d3e97aa2386917e96dffd0b750b79302c314fae1e0741dae05bfd404ccd54a3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6dd37bcc5fa8c3273532bfeef756e11f279121d2089197f5e6ff55e99580a00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bb56da23d00e0fddf638408a8a5ef2184587ff84e5e80c3580be43fb1266f92b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d150f005fe8358a5025410c0a421bed258d8e7697089f931565ba50809c941bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://32af04e4f956e57936737084d0a395007655ead1604b242bb0acdb0cdad76828\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://65c677dbb15e97b511c129bc19e0e8fbe77978acf58e1a86c62180ecd07bbf14\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c3cfa3e4e0a868699caf5972c9b61f5597ffd222c448774db4045da4b42766da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.132833 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f1b1bcec-5a4f-4926-a164-54a32764e774\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://96aab2c3c7e52336e6b6347587ac57d218cd6a907e8eae7e3966551a9625250f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24048ba5e57996d5d6a840c7964164d78205174af9159973abc3695050a71048\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0562cbb4cc18dca87848b015a5a5e54a764aeeecb2101e41626dc06364fe85d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bd9b53d998efce03feef151dffa8271f83e23564d11efa6b722d63d4a8d90336\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.150434 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.161567 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7142d4cf-9f77-4d00-be33-0198a4cb84d1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://48f4430140c0067a3d999fab6eb020d2701420bd61ff0cb77f8eb26eea27ad45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8rz8v\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-b82dg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.171674 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-xpn28" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"81f30c43-dacf-407f-adb3-bdbae866485f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dh6kd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:58Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-xpn28\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.186301 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vtrls" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d167a938-f26f-449c-9ee6-bd7247ac68e4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:52Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b883abcf28fab77a5f6dd07c7a30bdc571bbd220fa9b5f36a674d6894578e5a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30e33a8ada484ccbd90f9cb3b8f83111d8684deb34387db897c7a51f3b437c15\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ef36245a2a9b798689b1e1ef0c7a2d9983e16f7ccc6f3bd946762a5188e7cf07\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://105c0970dc8f0029418af98662dca7c80146ab138f147584f5fe3e0b44939305\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c61b420b66fbce26669cbb53e2973a3e6805dd9c8a391189c3580e652204ccd0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://21370b828cb7274f4ce7144211297dc94b4988fed87db483b0b7e3f1624bb3da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://42a5e6aaad68e7de43d204a3cd26786473a85e285aebb328876d8e47be493833\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:50Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pqmcs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vtrls\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.187299 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.187330 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.187549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.187585 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.187597 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.203677 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a4d98c7-0970-4ab3-86a6-40c61350f745\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-25T09:49:41Z\\\",\\\"message\\\":\\\"rc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:49:41Z is after 2025-08-24T17:21:41Z]\\\\nI1125 09:49:41.821139 6955 transact.go:42] Configuring OVN: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:49:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xtqrw\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:44Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-4pcsz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.215135 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27b20208-edb8-4717-876c-ce1991660bc8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:58Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2119adfe9ad0c255f10560023cd067a63f2099cad58aabedfcd1ec1dcc3b682b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ed832322274a1e41727f2e4066889a38edccf889cdca5afff7d092ca9e54b2c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-544bk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:57Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-vbg2l\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.231498 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b76756dc-1402-4054-bff2-e76ba1281b6f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:49:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-25T09:48:43Z\\\",\\\"message\\\":\\\"lling back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1125 09:48:37.753693 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1125 09:48:37.755445 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-952826183/tls.crt::/tmp/serving-cert-952826183/tls.key\\\\\\\"\\\\nI1125 09:48:43.378319 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1125 09:48:43.383921 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1125 09:48:43.383969 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1125 09:48:43.383994 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1125 09:48:43.384001 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1125 09:48:43.391832 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1125 09:48:43.391871 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391878 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1125 09:48:43.391883 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1125 09:48:43.391887 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1125 09:48:43.391890 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1125 09:48:43.391894 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1125 09:48:43.392263 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1125 09:48:43.394348 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:27Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:26Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-25T09:48:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.245490 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c87acdba-75b3-4150-881d-7acdf123516d\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:26Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a1beff832aac3af2c26897bdb775d5fe7b5039fcc7c1feb7a6fead0123841de3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://53aa421c16cc0add586b070927f3b9044f91a48a934aabfe61a9154c95083f3b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fba87b48ac2ba62a73d2b21416d33626f480bfbeb16f39162ec7cbc31032228b\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:24Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.258571 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c8c82aef5cd9fb8340d5b8bebaaad549c398f1fbfd4c884bf9672d71f842d5d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44c3e947b8fcdd8b9c73918736e4c711d079e191fdedc1a3f89523922c347392\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.272272 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b0daaa0fa8f7c72fe15f6cdb076d25a5d83e77cc2b86f5cb714fce44a1dd728c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.282745 4926 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-pfggm" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"de954742-8a79-473e-bcb4-537799771043\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-25T09:48:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0b14e71f3b4f1c2630f57cd65439641d36e4d35b3d914946e50006958d07d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-25T09:48:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zrkks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-25T09:48:47Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-pfggm\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:04Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.289635 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.289668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.289679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.289696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.289706 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.393025 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.393116 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.393130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.393151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.393163 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.496228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.496284 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.496294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.496313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.496324 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.600241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.600298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.600308 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.600325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.600338 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.703618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.703678 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.703688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.703706 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.703740 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.806385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.806427 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.806462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.806483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.806495 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.908796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.908845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.908858 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.908878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:04 crc kubenswrapper[4926]: I1125 09:50:04.908889 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:04Z","lastTransitionTime":"2025-11-25T09:50:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.012047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.012141 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.012158 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.012178 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.012220 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.114489 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.114532 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.114542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.114563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.114575 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.217879 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.217911 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.217919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.217935 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.217945 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.320803 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.320874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.320886 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.320908 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.320921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.424147 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.424201 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.424215 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.424247 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.424259 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.526835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.526878 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.526888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.526905 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.526915 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.630136 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.630182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.630193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.630209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.630219 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.733099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.733160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.733170 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.733190 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.733200 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.836737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.836797 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.836808 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.836833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.836851 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.939220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.939286 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.939297 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.939320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:05 crc kubenswrapper[4926]: I1125 09:50:05.939332 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:05Z","lastTransitionTime":"2025-11-25T09:50:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.012028 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.012293 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.012466 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:06 crc kubenswrapper[4926]: E1125 09:50:06.012598 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.012647 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:06 crc kubenswrapper[4926]: E1125 09:50:06.012750 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:06 crc kubenswrapper[4926]: E1125 09:50:06.012858 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:06 crc kubenswrapper[4926]: E1125 09:50:06.013052 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.041689 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.041748 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.041757 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.041775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.041785 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.143845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.143900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.143913 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.143955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.143969 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.247373 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.247425 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.247434 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.247452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.247462 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.350728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.350785 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.350798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.350839 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.350854 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.454420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.454452 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.454462 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.454479 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.454489 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.557676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.557716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.557726 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.557743 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.557758 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.662712 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.662764 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.662778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.662801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.662824 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.765608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.765650 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.765661 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.765680 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.765692 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.869003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.869070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.869085 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.869109 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.869125 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.972395 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.972473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.972490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.972512 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:06 crc kubenswrapper[4926]: I1125 09:50:06.972534 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:06Z","lastTransitionTime":"2025-11-25T09:50:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.076126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.076193 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.076214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.076238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.076253 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.179801 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.179855 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.179865 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.179882 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.179894 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.282477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.282526 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.282538 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.282560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.282574 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.385248 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.385295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.385310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.385335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.385351 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.488179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.488213 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.488225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.488245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.488256 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.591097 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.591197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.591210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.591229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.591239 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.694165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.694238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.694255 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.694276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.694289 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.796524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.796569 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.796581 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.796600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.796613 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.899505 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.899560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.899570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.899588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:07 crc kubenswrapper[4926]: I1125 09:50:07.899604 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:07Z","lastTransitionTime":"2025-11-25T09:50:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.002774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.002833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.002850 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.002875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.002892 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.012413 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.012415 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.012534 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:08 crc kubenswrapper[4926]: E1125 09:50:08.012679 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.012768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:08 crc kubenswrapper[4926]: E1125 09:50:08.012900 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:08 crc kubenswrapper[4926]: E1125 09:50:08.013055 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:08 crc kubenswrapper[4926]: E1125 09:50:08.013215 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.014064 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:50:08 crc kubenswrapper[4926]: E1125 09:50:08.014808 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-4pcsz_openshift-ovn-kubernetes(9a4d98c7-0970-4ab3-86a6-40c61350f745)\"" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.106364 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.106411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.106423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.106445 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.106459 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.210003 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.210048 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.210100 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.210120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.210133 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.313520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.313563 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.313583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.313603 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.313617 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.416682 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.416721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.416731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.416747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.416757 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.520114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.520160 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.520171 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.520190 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.520200 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.622821 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.622869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.622880 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.622899 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.622913 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.725532 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.725588 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.725601 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.725622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.725637 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.828995 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.829035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.829043 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.829059 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.829069 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.932151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.932211 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.932221 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.932237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:08 crc kubenswrapper[4926]: I1125 09:50:08.932245 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:08Z","lastTransitionTime":"2025-11-25T09:50:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.034747 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.034794 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.034805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.034828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.034843 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.137324 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.137367 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.137378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.137396 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.137408 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.239673 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.239716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.239728 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.239746 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.239758 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.342320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.342425 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.342439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.342456 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.342466 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.448366 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.448438 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.448453 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.448476 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.448489 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.553149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.553271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.553295 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.553328 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.553348 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.657012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.657073 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.657083 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.657102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.657112 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.759762 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.759815 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.759829 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.759852 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.759867 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.862670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.862707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.862716 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.862733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.862742 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.965012 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.965057 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.965088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.965105 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:09 crc kubenswrapper[4926]: I1125 09:50:09.965115 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:09Z","lastTransitionTime":"2025-11-25T09:50:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.011806 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.011869 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.011814 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.012016 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:10 crc kubenswrapper[4926]: E1125 09:50:10.012170 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:10 crc kubenswrapper[4926]: E1125 09:50:10.012281 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:10 crc kubenswrapper[4926]: E1125 09:50:10.012354 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:10 crc kubenswrapper[4926]: E1125 09:50:10.012445 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.068121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.068182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.068192 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.068210 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.068221 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.170652 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.170714 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.170730 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.170754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.170768 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.273195 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.273243 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.273254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.273272 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.273285 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.376702 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.376749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.376761 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.376777 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.376791 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.480184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.480246 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.480257 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.480276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.480287 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.585300 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.585350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.585361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.585380 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.585393 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.687618 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.687668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.687679 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.687698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.687710 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.791120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.791194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.791214 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.791239 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.791249 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.893671 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.893745 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.893768 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.893798 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.893822 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.996833 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.996903 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.996927 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.997002 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:10 crc kubenswrapper[4926]: I1125 09:50:10.997035 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:10Z","lastTransitionTime":"2025-11-25T09:50:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.099173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.099270 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.099298 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.099331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.099355 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.202394 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.202449 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.202461 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.202486 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.202502 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.305675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.305763 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.305778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.305802 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.305818 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.409047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.409099 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.409112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.409134 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.409152 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.512031 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.512077 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.512090 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.512112 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.512124 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.614608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.614662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.614674 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.614694 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.614710 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.717758 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.717805 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.717816 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.717835 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.717846 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.820725 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.820778 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.820790 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.820806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.820816 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.923062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.923104 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.923113 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.923130 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:11 crc kubenswrapper[4926]: I1125 09:50:11.923139 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:11Z","lastTransitionTime":"2025-11-25T09:50:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.011541 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.011612 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.011621 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.011562 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.011728 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.011790 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.011924 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.012098 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.025483 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.025525 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.025540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.025562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.025578 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.129110 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.129154 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.129165 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.129184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.129226 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.233657 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.233710 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.233720 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.233740 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.233752 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.307041 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.307093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.307107 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.307126 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.307140 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.321166 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.325485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.325532 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.325542 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.325560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.325570 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.341689 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.345834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.345864 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.345872 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.345887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.345896 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.359112 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.363069 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.363108 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.363118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.363145 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.363156 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.377502 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.382035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.382093 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.382117 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.382151 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.382176 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.398460 4926 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-25T09:50:12Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ae35d4ee-cf02-4f2c-87da-f07de9ee360a\\\",\\\"systemUUID\\\":\\\"612af39e-7a36-4010-be71-d81687d215a9\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-25T09:50:12Z is after 2025-08-24T17:21:41Z" Nov 25 09:50:12 crc kubenswrapper[4926]: E1125 09:50:12.398627 4926 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.400804 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.400848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.400868 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.400892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.400902 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.504281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.504350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.504368 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.504390 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.504431 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.607309 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.607391 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.607410 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.607430 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.607441 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.710196 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.710241 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.710256 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.710277 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.710293 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.813549 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.813600 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.813612 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.813632 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.813645 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.916533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.916606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.916620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.916639 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:12 crc kubenswrapper[4926]: I1125 09:50:12.916650 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:12Z","lastTransitionTime":"2025-11-25T09:50:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.019229 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.019283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.019294 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.019317 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.019332 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.122864 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.122919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.122967 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.122989 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.123003 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.225558 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.225620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.225642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.225668 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.225686 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.328414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.328459 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.328473 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.328506 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.328516 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.431900 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.431968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.431983 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.432011 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.432026 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.535326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.535456 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.535471 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.535499 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.535559 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.638800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.638842 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.638851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.638869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.638881 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.742258 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.742754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.743029 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.743258 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.743456 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.846609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.846648 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.846659 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.846676 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.846686 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.949184 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.949235 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.949245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.949263 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:13 crc kubenswrapper[4926]: I1125 09:50:13.949272 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:13Z","lastTransitionTime":"2025-11-25T09:50:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.012148 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.012229 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:14 crc kubenswrapper[4926]: E1125 09:50:14.012341 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.012167 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.012517 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:14 crc kubenswrapper[4926]: E1125 09:50:14.012573 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:14 crc kubenswrapper[4926]: E1125 09:50:14.012625 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:14 crc kubenswrapper[4926]: E1125 09:50:14.012696 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.054647 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.054687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.054696 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.054731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.054742 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.059183 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=90.059154383 podStartE2EDuration="1m30.059154383s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.041299921 +0000 UTC m=+110.725256368" watchObservedRunningTime="2025-11-25 09:50:14.059154383 +0000 UTC m=+110.743110820" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.075482 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=89.075458661 podStartE2EDuration="1m29.075458661s" podCreationTimestamp="2025-11-25 09:48:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.059428831 +0000 UTC m=+110.743385258" watchObservedRunningTime="2025-11-25 09:50:14.075458661 +0000 UTC m=+110.759415088" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.113868 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-pfggm" podStartSLOduration=90.113849902 podStartE2EDuration="1m30.113849902s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.098913194 +0000 UTC m=+110.782869621" watchObservedRunningTime="2025-11-25 09:50:14.113849902 +0000 UTC m=+110.797806329" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.114304 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-l78n4" podStartSLOduration=90.114295415 podStartE2EDuration="1m30.114295415s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.113390189 +0000 UTC m=+110.797346606" watchObservedRunningTime="2025-11-25 09:50:14.114295415 +0000 UTC m=+110.798251842" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.125263 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.125242619 podStartE2EDuration="31.125242619s" podCreationTimestamp="2025-11-25 09:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.12492397 +0000 UTC m=+110.808880397" watchObservedRunningTime="2025-11-25 09:50:14.125242619 +0000 UTC m=+110.809199046" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.158895 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.158930 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.158955 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.158972 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.158985 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.239345 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-58wwc" podStartSLOduration=90.23931701 podStartE2EDuration="1m30.23931701s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.182794339 +0000 UTC m=+110.866750766" watchObservedRunningTime="2025-11-25 09:50:14.23931701 +0000 UTC m=+110.923273447" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.239693 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=87.239687291 podStartE2EDuration="1m27.239687291s" podCreationTimestamp="2025-11-25 09:48:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.23651639 +0000 UTC m=+110.920472817" watchObservedRunningTime="2025-11-25 09:50:14.239687291 +0000 UTC m=+110.923643718" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.253140 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.253118566 podStartE2EDuration="56.253118566s" podCreationTimestamp="2025-11-25 09:49:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.252505479 +0000 UTC m=+110.936461906" watchObservedRunningTime="2025-11-25 09:50:14.253118566 +0000 UTC m=+110.937074993" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.261739 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.261786 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.261799 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.261821 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.261837 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.305366 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podStartSLOduration=90.305335514 podStartE2EDuration="1m30.305335514s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.289978763 +0000 UTC m=+110.973935190" watchObservedRunningTime="2025-11-25 09:50:14.305335514 +0000 UTC m=+110.989291941" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.321303 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vtrls" podStartSLOduration=90.321281581 podStartE2EDuration="1m30.321281581s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.320771656 +0000 UTC m=+111.004728083" watchObservedRunningTime="2025-11-25 09:50:14.321281581 +0000 UTC m=+111.005238008" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.364191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.364224 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.364233 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.364250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.364259 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.466930 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.466994 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.467004 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.467026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.467041 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.570714 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.570776 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.570793 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.570819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.570839 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.677067 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.677169 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.677212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.677238 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.677286 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.781142 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.781212 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.781228 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.781251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.781265 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.883984 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.884047 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.884058 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.884079 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.884091 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.987102 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.987150 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.987162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.987182 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:14 crc kubenswrapper[4926]: I1125 09:50:14.987194 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:14Z","lastTransitionTime":"2025-11-25T09:50:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.090749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.090789 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.090807 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.090828 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.090850 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.193121 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.193174 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.193266 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.193288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.193301 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.296245 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.296292 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.296302 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.296321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.296334 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.399149 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.399188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.399198 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.399216 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.399226 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.501732 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.501810 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.501837 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.501916 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.501978 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.604370 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.604414 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.604422 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.604439 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.604450 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.706749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.706874 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.706887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.706907 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.706921 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.809614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.809660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.809670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.809687 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.809698 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.912642 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.912684 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.912698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.912715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:15 crc kubenswrapper[4926]: I1125 09:50:15.912729 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:15Z","lastTransitionTime":"2025-11-25T09:50:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.012177 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.012231 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.012207 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:16 crc kubenswrapper[4926]: E1125 09:50:16.012349 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.012407 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:16 crc kubenswrapper[4926]: E1125 09:50:16.012450 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:16 crc kubenswrapper[4926]: E1125 09:50:16.012594 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:16 crc kubenswrapper[4926]: E1125 09:50:16.012672 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.014856 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.014898 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.014914 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.014932 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.014957 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.117884 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.117953 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.117968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.117988 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.118000 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.220800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.220851 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.220867 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.220887 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.220901 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.324333 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.324378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.324387 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.324408 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.324419 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.427420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.427477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.427490 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.427513 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.427527 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.530423 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.530470 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.530480 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.530495 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.530504 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.633622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.633688 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.633701 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.633722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.633734 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.735686 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.735733 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.735749 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.735775 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.735788 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.839570 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.839624 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.839638 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.839662 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.839676 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.941840 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.941888 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.941901 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.941919 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:16 crc kubenswrapper[4926]: I1125 09:50:16.941946 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:16Z","lastTransitionTime":"2025-11-25T09:50:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.045144 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.045191 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.045204 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.045223 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.045236 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.148018 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.148061 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.148070 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.148086 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.148096 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.251320 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.251376 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.251389 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.251412 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.251425 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.360475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.360543 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.360562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.360593 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.360611 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.463475 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.463520 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.463529 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.463547 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.463557 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.566555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.566606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.566615 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.566875 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.566900 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.669540 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.669596 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.669608 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.669627 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.669639 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.772693 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.772741 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.772754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.772772 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.772785 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.875614 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.875660 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.875675 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.875698 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.875715 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.978871 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.978920 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.978931 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.978968 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:17 crc kubenswrapper[4926]: I1125 09:50:17.978980 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:17Z","lastTransitionTime":"2025-11-25T09:50:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.011749 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.011853 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:18 crc kubenswrapper[4926]: E1125 09:50:18.011930 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:18 crc kubenswrapper[4926]: E1125 09:50:18.012086 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.012151 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:18 crc kubenswrapper[4926]: E1125 09:50:18.012195 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.012225 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:18 crc kubenswrapper[4926]: E1125 09:50:18.012298 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.082036 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.082084 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.082094 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.082111 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.082123 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.185120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.185173 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.185186 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.185207 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.185222 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.288285 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.288331 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.288342 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.288361 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.288374 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.390754 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.390806 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.390820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.390839 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.390854 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.493663 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.493699 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.493707 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.493722 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.493732 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.597110 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.597164 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.597174 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.597194 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.597209 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.611740 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/1.log" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.612327 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/0.log" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.612499 4926 generic.go:334] "Generic (PLEG): container finished" podID="cab27fcb-cde4-4879-8fad-79951474d7b3" containerID="1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54" exitCode=1 Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.612536 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerDied","Data":"1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.612574 4926 scope.go:117] "RemoveContainer" containerID="53d2b54abbe356432b4acfd18a6f385dd3896d89d4591f4e7448f329033113ff" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.613097 4926 scope.go:117] "RemoveContainer" containerID="1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54" Nov 25 09:50:18 crc kubenswrapper[4926]: E1125 09:50:18.613641 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-l78n4_openshift-multus(cab27fcb-cde4-4879-8fad-79951474d7b3)\"" pod="openshift-multus/multus-l78n4" podUID="cab27fcb-cde4-4879-8fad-79951474d7b3" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.631978 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-vbg2l" podStartSLOduration=94.631928587 podStartE2EDuration="1m34.631928587s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:14.363119841 +0000 UTC m=+111.047076288" watchObservedRunningTime="2025-11-25 09:50:18.631928587 +0000 UTC m=+115.315885034" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.700053 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.700092 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.700101 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.700119 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.700130 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.803774 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.803819 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.803830 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.803849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.803861 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.907014 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.907095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.907123 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.907152 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:18 crc kubenswrapper[4926]: I1125 09:50:18.907170 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:18Z","lastTransitionTime":"2025-11-25T09:50:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.010035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.010087 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.010098 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.010114 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.010125 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.113352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.113400 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.113411 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.113431 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.113443 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.216651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.216715 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.216731 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.216752 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.216766 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.319179 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.319225 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.319237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.319254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.319266 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.421737 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.421809 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.421822 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.421845 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.421859 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.525153 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.525197 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.525283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.525306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.525332 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.618154 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/1.log" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.627800 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.627834 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.627848 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.627866 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.627879 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.731580 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.731651 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.731665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.731685 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.731699 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.841562 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.841670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.841691 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.841756 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.841784 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.945405 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.945443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.945463 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.945485 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:19 crc kubenswrapper[4926]: I1125 09:50:19.945495 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:19Z","lastTransitionTime":"2025-11-25T09:50:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.015393 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:20 crc kubenswrapper[4926]: E1125 09:50:20.015563 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.015651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:20 crc kubenswrapper[4926]: E1125 09:50:20.015719 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.015779 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:20 crc kubenswrapper[4926]: E1125 09:50:20.015854 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.015913 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:20 crc kubenswrapper[4926]: E1125 09:50:20.016000 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.048310 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.048369 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.048381 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.048401 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.048413 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.151226 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.151269 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.151281 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.151338 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.151353 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.254120 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.254208 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.254220 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.254237 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.254248 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.357509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.357560 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.357578 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.357607 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.357625 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.461288 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.461326 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.461335 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.461350 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.461361 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.565095 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.565162 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.565177 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.565201 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.565217 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.667796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.667854 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.667869 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.667892 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.667907 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.770884 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.771006 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.771026 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.771062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.771087 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.873929 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.874062 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.874088 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.874118 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.874140 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.977283 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.977325 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.977334 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.977352 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:20 crc kubenswrapper[4926]: I1125 09:50:20.977363 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:20Z","lastTransitionTime":"2025-11-25T09:50:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.080477 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.080521 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.080533 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.080555 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.080569 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.184443 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.184559 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.184584 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.184620 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.184646 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.288494 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.288546 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.288583 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.288606 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.288618 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.391209 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.391244 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.391254 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.391276 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.391292 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.494738 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.494796 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.494820 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.494849 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.494867 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.598670 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.598721 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.598736 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.598760 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.598772 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.702388 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.702463 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.702480 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.702509 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.702537 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.805289 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.805345 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.805358 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.805378 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.805391 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.909035 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.909125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.909157 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.909199 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:21 crc kubenswrapper[4926]: I1125 09:50:21.909227 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:21Z","lastTransitionTime":"2025-11-25T09:50:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.011329 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.011499 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.011584 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.011638 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:22 crc kubenswrapper[4926]: E1125 09:50:22.011527 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:22 crc kubenswrapper[4926]: E1125 09:50:22.011859 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:22 crc kubenswrapper[4926]: E1125 09:50:22.012023 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.012321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: E1125 09:50:22.012327 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.012351 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.012420 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.012524 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.012599 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.115561 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.115622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.115640 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.115665 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.115684 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.223552 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.223609 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.223622 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.223649 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.223663 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.326913 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.326973 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.326986 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.327125 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.327142 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.429250 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.429306 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.429321 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.429337 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.429349 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.532188 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.532236 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.532251 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.532271 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.532284 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.557313 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.557385 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.557399 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.557426 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.557441 4926 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-25T09:50:22Z","lastTransitionTime":"2025-11-25T09:50:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.605221 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw"] Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.605688 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.607610 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.607899 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.607956 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.608149 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.652508 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/731d2e7e-81c3-444b-8599-95c85f042744-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.652570 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/731d2e7e-81c3-444b-8599-95c85f042744-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.652596 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/731d2e7e-81c3-444b-8599-95c85f042744-service-ca\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.652611 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/731d2e7e-81c3-444b-8599-95c85f042744-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.652737 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/731d2e7e-81c3-444b-8599-95c85f042744-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/731d2e7e-81c3-444b-8599-95c85f042744-service-ca\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754143 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/731d2e7e-81c3-444b-8599-95c85f042744-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754201 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/731d2e7e-81c3-444b-8599-95c85f042744-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754351 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/731d2e7e-81c3-444b-8599-95c85f042744-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754294 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/731d2e7e-81c3-444b-8599-95c85f042744-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754430 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/731d2e7e-81c3-444b-8599-95c85f042744-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.754508 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/731d2e7e-81c3-444b-8599-95c85f042744-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.755975 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/731d2e7e-81c3-444b-8599-95c85f042744-service-ca\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.761364 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/731d2e7e-81c3-444b-8599-95c85f042744-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.771814 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/731d2e7e-81c3-444b-8599-95c85f042744-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-87jlw\" (UID: \"731d2e7e-81c3-444b-8599-95c85f042744\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:22 crc kubenswrapper[4926]: I1125 09:50:22.919982 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.012211 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.634321 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/3.log" Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.637743 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerStarted","Data":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.638391 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.639177 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" event={"ID":"731d2e7e-81c3-444b-8599-95c85f042744","Type":"ContainerStarted","Data":"389618bb9d8558e23965f993bdda753d7736a572a1cda93942c211424b2d771c"} Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.639215 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" event={"ID":"731d2e7e-81c3-444b-8599-95c85f042744","Type":"ContainerStarted","Data":"cf2af6a720e3cb2a533431b432a2dd088fc323f4920b37ad5627dc7ebcae1bbd"} Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.682835 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-87jlw" podStartSLOduration=99.682808153 podStartE2EDuration="1m39.682808153s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:23.681486985 +0000 UTC m=+120.365443412" watchObservedRunningTime="2025-11-25 09:50:23.682808153 +0000 UTC m=+120.366764580" Nov 25 09:50:23 crc kubenswrapper[4926]: I1125 09:50:23.683249 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podStartSLOduration=99.683240806 podStartE2EDuration="1m39.683240806s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:23.668090701 +0000 UTC m=+120.352047138" watchObservedRunningTime="2025-11-25 09:50:23.683240806 +0000 UTC m=+120.367197233" Nov 25 09:50:23 crc kubenswrapper[4926]: E1125 09:50:23.982495 4926 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 25 09:50:24 crc kubenswrapper[4926]: I1125 09:50:24.011537 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:24 crc kubenswrapper[4926]: I1125 09:50:24.011577 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:24 crc kubenswrapper[4926]: I1125 09:50:24.011554 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:24 crc kubenswrapper[4926]: E1125 09:50:24.012846 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:24 crc kubenswrapper[4926]: I1125 09:50:24.013157 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:24 crc kubenswrapper[4926]: E1125 09:50:24.013243 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:24 crc kubenswrapper[4926]: E1125 09:50:24.013399 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:24 crc kubenswrapper[4926]: E1125 09:50:24.013956 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:24 crc kubenswrapper[4926]: I1125 09:50:24.068322 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xpn28"] Nov 25 09:50:24 crc kubenswrapper[4926]: E1125 09:50:24.105351 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:50:24 crc kubenswrapper[4926]: I1125 09:50:24.643099 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:24 crc kubenswrapper[4926]: E1125 09:50:24.643258 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:26 crc kubenswrapper[4926]: I1125 09:50:26.011346 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:26 crc kubenswrapper[4926]: E1125 09:50:26.011473 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:26 crc kubenswrapper[4926]: I1125 09:50:26.011530 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:26 crc kubenswrapper[4926]: I1125 09:50:26.011568 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:26 crc kubenswrapper[4926]: E1125 09:50:26.011612 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:26 crc kubenswrapper[4926]: I1125 09:50:26.011346 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:26 crc kubenswrapper[4926]: E1125 09:50:26.011768 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:26 crc kubenswrapper[4926]: E1125 09:50:26.012109 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:28 crc kubenswrapper[4926]: I1125 09:50:28.012320 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:28 crc kubenswrapper[4926]: I1125 09:50:28.012391 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:28 crc kubenswrapper[4926]: I1125 09:50:28.012344 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:28 crc kubenswrapper[4926]: I1125 09:50:28.012319 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:28 crc kubenswrapper[4926]: E1125 09:50:28.012510 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:28 crc kubenswrapper[4926]: E1125 09:50:28.012852 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:28 crc kubenswrapper[4926]: E1125 09:50:28.012972 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:28 crc kubenswrapper[4926]: E1125 09:50:28.013043 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:29 crc kubenswrapper[4926]: E1125 09:50:29.106428 4926 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 09:50:30 crc kubenswrapper[4926]: I1125 09:50:30.012142 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:30 crc kubenswrapper[4926]: I1125 09:50:30.012193 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:30 crc kubenswrapper[4926]: E1125 09:50:30.012608 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:30 crc kubenswrapper[4926]: I1125 09:50:30.012278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:30 crc kubenswrapper[4926]: I1125 09:50:30.012249 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:30 crc kubenswrapper[4926]: E1125 09:50:30.012701 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:30 crc kubenswrapper[4926]: E1125 09:50:30.012850 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:30 crc kubenswrapper[4926]: E1125 09:50:30.012966 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.011986 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.011986 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:32 crc kubenswrapper[4926]: E1125 09:50:32.012193 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.012040 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.012018 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:32 crc kubenswrapper[4926]: E1125 09:50:32.012412 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:32 crc kubenswrapper[4926]: E1125 09:50:32.012456 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:32 crc kubenswrapper[4926]: E1125 09:50:32.012536 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.012986 4926 scope.go:117] "RemoveContainer" containerID="1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.671642 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/1.log" Nov 25 09:50:32 crc kubenswrapper[4926]: I1125 09:50:32.671726 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerStarted","Data":"e52d67a89b7079554fcf1c05dd1fccff682bba2a9a44c7601776d730e21d6fe6"} Nov 25 09:50:34 crc kubenswrapper[4926]: I1125 09:50:34.011606 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:34 crc kubenswrapper[4926]: I1125 09:50:34.011617 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:34 crc kubenswrapper[4926]: I1125 09:50:34.011628 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:34 crc kubenswrapper[4926]: I1125 09:50:34.011559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:34 crc kubenswrapper[4926]: E1125 09:50:34.013238 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-xpn28" podUID="81f30c43-dacf-407f-adb3-bdbae866485f" Nov 25 09:50:34 crc kubenswrapper[4926]: E1125 09:50:34.013482 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 25 09:50:34 crc kubenswrapper[4926]: E1125 09:50:34.013595 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 25 09:50:34 crc kubenswrapper[4926]: E1125 09:50:34.013411 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.011499 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.011507 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.011598 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.011648 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.015263 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.015785 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.015954 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.016834 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.017191 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 09:50:36 crc kubenswrapper[4926]: I1125 09:50:36.017230 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.114873 4926 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.155766 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.156405 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.158548 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sw9lq"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.159000 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.160819 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6djth"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.161632 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.162820 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.163014 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.163159 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.163202 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.165026 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.170666 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.171020 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.171169 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.175133 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.175193 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.175454 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.175712 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.175768 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.175925 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.180041 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.180041 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.180384 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.180438 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.180809 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.182213 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.182763 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.183322 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9pvq7"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.183912 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.185050 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-hfbwq"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.199234 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.205741 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v7qs6"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.206256 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.207030 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bd494"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.207331 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.208153 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.208693 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.209873 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.210089 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.210241 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.210423 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.210571 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.210663 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.211115 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.211450 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.211967 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.212025 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.212401 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.212747 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.212971 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.213228 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.213906 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.216313 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.216627 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.216848 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zltj"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.217155 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.217473 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.217676 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.218315 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.218583 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.218510 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.219149 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6g2vd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.219743 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.220354 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dnjv7"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.220871 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.221028 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-68sgp"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.221417 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.222894 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223284 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223389 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223453 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223497 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223506 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223566 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223594 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223601 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223697 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.223852 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.225345 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.225721 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.228562 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.229099 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.229467 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.229920 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.230083 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.230189 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.233195 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.233729 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.234409 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.234678 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.240297 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.271490 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.272893 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-bngtn"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.273542 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.273963 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.274503 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.275587 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283110 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-config\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283151 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283183 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77kbb\" (UniqueName: \"kubernetes.io/projected/1413207c-fe4e-4f65-8d8f-605ba4474b63-kube-api-access-77kbb\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283200 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-audit-policies\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283218 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-config\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283252 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c67403f1-0cf5-4b25-93dd-36253a6135e2-config\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283267 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c67403f1-0cf5-4b25-93dd-36253a6135e2-trusted-ca\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283309 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltmhq\" (UniqueName: \"kubernetes.io/projected/60183186-5dd7-4632-96ef-2e87fb0c821a-kube-api-access-ltmhq\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-trusted-ca-bundle\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283344 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-encryption-config\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283363 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f3051c-91c1-44bc-a990-4f0a1165b4d2-serving-cert\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283385 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-oauth-config\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283402 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-service-ca\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283421 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thpx7\" (UniqueName: \"kubernetes.io/projected/c67403f1-0cf5-4b25-93dd-36253a6135e2-kube-api-access-thpx7\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283440 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-console-config\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-client-ca\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283482 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwd54\" (UniqueName: \"kubernetes.io/projected/6a6a3048-e831-4418-baf5-7bca98a35cef-kube-api-access-jwd54\") pod \"cluster-samples-operator-665b6dd947-vcbkn\" (UID: \"6a6a3048-e831-4418-baf5-7bca98a35cef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.283510 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-serving-cert\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286411 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6znh4\" (UniqueName: \"kubernetes.io/projected/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-kube-api-access-6znh4\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286490 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7x6j\" (UniqueName: \"kubernetes.io/projected/65f3051c-91c1-44bc-a990-4f0a1165b4d2-kube-api-access-z7x6j\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286571 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-images\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286579 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a6a3048-e831-4418-baf5-7bca98a35cef-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vcbkn\" (UID: \"6a6a3048-e831-4418-baf5-7bca98a35cef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286638 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-etcd-client\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286701 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c67403f1-0cf5-4b25-93dd-36253a6135e2-serving-cert\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286750 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286774 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-oauth-serving-cert\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286802 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1413207c-fe4e-4f65-8d8f-605ba4474b63-audit-dir\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286822 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-serving-cert\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.286847 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.287003 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.287177 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.287725 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288001 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288108 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288174 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288244 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288336 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288373 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288457 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288547 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288667 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288777 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.288925 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.289072 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.289197 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.289313 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.289437 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.289585 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.289850 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290081 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290252 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290275 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290371 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290432 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290488 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290626 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290741 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290783 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.290862 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291006 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.293386 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sw9lq"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291094 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.299117 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.299210 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291148 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291199 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291252 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291305 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291350 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291396 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291433 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291471 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291507 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291546 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291582 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291622 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291682 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291827 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.291864 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.293362 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.293439 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.293576 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.299821 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.298305 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.298643 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.298710 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.300621 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.302669 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.306251 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.307763 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.310354 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.310971 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.311006 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.311467 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.311604 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.314993 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.315020 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nn4ms"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.315759 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.316310 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.316402 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.319231 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.319362 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6xzbk"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.320324 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.326171 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.330482 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.331494 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.332328 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht99z"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.332915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.333124 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mztxj"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.333736 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.336783 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.337708 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.339049 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.347188 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6djth"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.348870 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.352018 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.369312 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hw4f2"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.369973 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.370726 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.371213 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v7qs6"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.371244 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9pvq7"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.371261 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.371362 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.374727 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bd494"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.375540 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.379749 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.386009 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387419 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387470 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-config\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387643 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387734 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-audit-policies\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387808 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-77kbb\" (UniqueName: \"kubernetes.io/projected/1413207c-fe4e-4f65-8d8f-605ba4474b63-kube-api-access-77kbb\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387891 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-config\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.387999 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c67403f1-0cf5-4b25-93dd-36253a6135e2-config\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388083 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c67403f1-0cf5-4b25-93dd-36253a6135e2-trusted-ca\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltmhq\" (UniqueName: \"kubernetes.io/projected/60183186-5dd7-4632-96ef-2e87fb0c821a-kube-api-access-ltmhq\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1200b8de-e5c0-4315-a8bc-0284c5963a76-config\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388319 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwvnq\" (UniqueName: \"kubernetes.io/projected/f2e3a659-20d7-47a9-9f7d-256a88c0e315-kube-api-access-fwvnq\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388401 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-trusted-ca-bundle\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388479 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-encryption-config\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388580 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f3051c-91c1-44bc-a990-4f0a1165b4d2-serving-cert\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388667 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-oauth-config\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388753 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thpx7\" (UniqueName: \"kubernetes.io/projected/c67403f1-0cf5-4b25-93dd-36253a6135e2-kube-api-access-thpx7\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388823 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-console-config\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-service-ca\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-config\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.389057 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-client-ca\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396173 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwd54\" (UniqueName: \"kubernetes.io/projected/6a6a3048-e831-4418-baf5-7bca98a35cef-kube-api-access-jwd54\") pod \"cluster-samples-operator-665b6dd947-vcbkn\" (UID: \"6a6a3048-e831-4418-baf5-7bca98a35cef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396360 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-serving-cert\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396454 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6znh4\" (UniqueName: \"kubernetes.io/projected/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-kube-api-access-6znh4\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396564 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-images\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7x6j\" (UniqueName: \"kubernetes.io/projected/65f3051c-91c1-44bc-a990-4f0a1165b4d2-kube-api-access-z7x6j\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.390352 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-console-config\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.388355 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.389756 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c67403f1-0cf5-4b25-93dd-36253a6135e2-trusted-ca\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.392485 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-config\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.393931 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c67403f1-0cf5-4b25-93dd-36253a6135e2-config\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.391814 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-trusted-ca-bundle\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.395737 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-encryption-config\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.395745 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f3051c-91c1-44bc-a990-4f0a1165b4d2-serving-cert\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.390783 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-service-ca\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.391397 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-audit-policies\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.390074 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.390369 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-client-ca\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2e3a659-20d7-47a9-9f7d-256a88c0e315-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396842 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6g2vd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.396983 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-oauth-config\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.397010 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.397029 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nn4ms"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.397047 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.397824 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-images\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.398742 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2dvw\" (UniqueName: \"kubernetes.io/projected/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-kube-api-access-q2dvw\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.398876 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a6a3048-e831-4418-baf5-7bca98a35cef-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vcbkn\" (UID: \"6a6a3048-e831-4418-baf5-7bca98a35cef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.398976 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-etcd-client\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399139 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-proxy-tls\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399221 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1200b8de-e5c0-4315-a8bc-0284c5963a76-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c67403f1-0cf5-4b25-93dd-36253a6135e2-serving-cert\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399501 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399598 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2e3a659-20d7-47a9-9f7d-256a88c0e315-proxy-tls\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1413207c-fe4e-4f65-8d8f-605ba4474b63-audit-dir\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399764 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-serving-cert\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399843 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-oauth-serving-cert\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.399921 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1200b8de-e5c0-4315-a8bc-0284c5963a76-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.400027 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.398824 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-images\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.401591 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-serving-cert\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.401730 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1413207c-fe4e-4f65-8d8f-605ba4474b63-audit-dir\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.398685 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.401805 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.401825 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zltj"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.402352 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.402450 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6a6a3048-e831-4418-baf5-7bca98a35cef-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-vcbkn\" (UID: \"6a6a3048-e831-4418-baf5-7bca98a35cef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.402599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1413207c-fe4e-4f65-8d8f-605ba4474b63-etcd-client\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.402649 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-oauth-serving-cert\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.403036 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.404002 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-serving-cert\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.404444 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.404563 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.405858 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.405396 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.406839 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht99z"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.408282 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1413207c-fe4e-4f65-8d8f-605ba4474b63-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.408476 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-brh52"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.408796 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.409348 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.411645 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-stkz9"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.412638 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.412748 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.413270 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.415023 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.415296 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c67403f1-0cf5-4b25-93dd-36253a6135e2-serving-cert\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.416050 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.417316 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dnjv7"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.418207 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-68sgp"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.419638 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.419931 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.421083 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-hfbwq"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.422338 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.423343 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.424424 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mztxj"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.426385 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hw4f2"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.427248 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6xzbk"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.428785 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-stkz9"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.430050 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.430247 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rvspl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.431260 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.431331 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rvspl"] Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.439528 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.459061 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.479529 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.499459 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500768 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-proxy-tls\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500797 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500818 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1200b8de-e5c0-4315-a8bc-0284c5963a76-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500848 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1200b8de-e5c0-4315-a8bc-0284c5963a76-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500869 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2e3a659-20d7-47a9-9f7d-256a88c0e315-proxy-tls\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1200b8de-e5c0-4315-a8bc-0284c5963a76-config\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.500991 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwvnq\" (UniqueName: \"kubernetes.io/projected/f2e3a659-20d7-47a9-9f7d-256a88c0e315-kube-api-access-fwvnq\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.501064 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-images\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.501087 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2e3a659-20d7-47a9-9f7d-256a88c0e315-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.501109 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2dvw\" (UniqueName: \"kubernetes.io/projected/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-kube-api-access-q2dvw\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.501789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1200b8de-e5c0-4315-a8bc-0284c5963a76-config\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.502120 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.502549 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f2e3a659-20d7-47a9-9f7d-256a88c0e315-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.505036 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f2e3a659-20d7-47a9-9f7d-256a88c0e315-proxy-tls\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.506413 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1200b8de-e5c0-4315-a8bc-0284c5963a76-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.518769 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.539843 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.560480 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.579309 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.599005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.618568 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.638672 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.659316 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.680221 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.698912 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.720009 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.739246 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.759923 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.779740 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.798880 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.819273 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.837869 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.858990 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.879902 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.899291 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.939253 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.942829 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-images\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.959102 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.966581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-proxy-tls\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:43 crc kubenswrapper[4926]: I1125 09:50:43.988673 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.019296 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.038750 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.059424 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.078685 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.100002 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.118831 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.139635 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.159315 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.179361 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.198506 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.227022 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.238603 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.298427 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.298680 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.299645 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.318619 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.337567 4926 request.go:700] Waited for 1.016929893s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.339194 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.359244 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.379491 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.399110 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.419891 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.439886 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.469079 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.478358 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.499820 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.519290 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.539988 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.560271 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.579905 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.604355 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.619158 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.638713 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.659227 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.678009 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.704628 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.719131 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.739967 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.759773 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.779022 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.799368 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.819100 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.839922 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.858573 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.878123 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.899431 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.920319 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.939562 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.958692 4926 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.979209 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 09:50:44 crc kubenswrapper[4926]: I1125 09:50:44.998597 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.034693 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltmhq\" (UniqueName: \"kubernetes.io/projected/60183186-5dd7-4632-96ef-2e87fb0c821a-kube-api-access-ltmhq\") pod \"console-f9d7485db-hfbwq\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.054463 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-77kbb\" (UniqueName: \"kubernetes.io/projected/1413207c-fe4e-4f65-8d8f-605ba4474b63-kube-api-access-77kbb\") pod \"apiserver-7bbb656c7d-t9spw\" (UID: \"1413207c-fe4e-4f65-8d8f-605ba4474b63\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.074839 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thpx7\" (UniqueName: \"kubernetes.io/projected/c67403f1-0cf5-4b25-93dd-36253a6135e2-kube-api-access-thpx7\") pod \"console-operator-58897d9998-6djth\" (UID: \"c67403f1-0cf5-4b25-93dd-36253a6135e2\") " pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.097705 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwd54\" (UniqueName: \"kubernetes.io/projected/6a6a3048-e831-4418-baf5-7bca98a35cef-kube-api-access-jwd54\") pod \"cluster-samples-operator-665b6dd947-vcbkn\" (UID: \"6a6a3048-e831-4418-baf5-7bca98a35cef\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.116721 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6znh4\" (UniqueName: \"kubernetes.io/projected/ec3f0e93-5599-45c0-b6d0-5e16be0df94c-kube-api-access-6znh4\") pod \"machine-api-operator-5694c8668f-9pvq7\" (UID: \"ec3f0e93-5599-45c0-b6d0-5e16be0df94c\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.119309 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.138342 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7x6j\" (UniqueName: \"kubernetes.io/projected/65f3051c-91c1-44bc-a990-4f0a1165b4d2-kube-api-access-z7x6j\") pod \"controller-manager-879f6c89f-sw9lq\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.139237 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.159738 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.179900 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.200443 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.219254 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.239344 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.260042 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.277180 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.279607 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.299533 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.314228 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.319181 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.325619 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.337052 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-hfbwq"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.355982 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2dvw\" (UniqueName: \"kubernetes.io/projected/12d97152-164d-4a5b-b3a0-ebfdb983d2d4-kube-api-access-q2dvw\") pod \"machine-config-operator-74547568cd-tqcdd\" (UID: \"12d97152-164d-4a5b-b3a0-ebfdb983d2d4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.356914 4926 request.go:700] Waited for 1.855018349s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/serviceaccounts/machine-config-controller/token Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.366437 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.402262 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.402754 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.408878 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwvnq\" (UniqueName: \"kubernetes.io/projected/f2e3a659-20d7-47a9-9f7d-256a88c0e315-kube-api-access-fwvnq\") pod \"machine-config-controller-84d6567774-ld9cl\" (UID: \"f2e3a659-20d7-47a9-9f7d-256a88c0e315\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.429345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.429832 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cab47b19-b54c-4498-90cf-2281376ffa72-webhook-cert\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.429882 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-etcd-serving-ca\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.429909 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba95982d-4bef-422c-99d9-16fa4f0160ed-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430052 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-stats-auth\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430078 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a31db46c-756f-4dff-bf2a-24c900683bb8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430133 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430152 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545a5916-91fb-4243-b825-fc09e352c3d5-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430168 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-auth-proxy-config\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430211 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kt6lg\" (UniqueName: \"kubernetes.io/projected/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-kube-api-access-kt6lg\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430230 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hvz4\" (UniqueName: \"kubernetes.io/projected/b3716ff0-7157-4406-be92-89893b24a9ef-kube-api-access-8hvz4\") pod \"multus-admission-controller-857f4d67dd-68sgp\" (UID: \"b3716ff0-7157-4406-be92-89893b24a9ef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430301 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdlmb\" (UniqueName: \"kubernetes.io/projected/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-kube-api-access-xdlmb\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430320 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnrsk\" (UniqueName: \"kubernetes.io/projected/cab47b19-b54c-4498-90cf-2281376ffa72-kube-api-access-bnrsk\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430369 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-config\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430391 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545a5916-91fb-4243-b825-fc09e352c3d5-trusted-ca\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430411 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430454 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a31db46c-756f-4dff-bf2a-24c900683bb8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430480 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-config\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430533 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a29e45e-3063-4a6d-b215-0765aa2df10a-service-ca-bundle\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430553 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430572 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430615 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-config\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430634 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wvpn\" (UniqueName: \"kubernetes.io/projected/146577d7-d2c0-4bf4-9d6b-236967454f83-kube-api-access-4wvpn\") pod \"control-plane-machine-set-operator-78cbb6b69f-xnjf9\" (UID: \"146577d7-d2c0-4bf4-9d6b-236967454f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430652 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nvql\" (UniqueName: \"kubernetes.io/projected/545a5916-91fb-4243-b825-fc09e352c3d5-kube-api-access-5nvql\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-encryption-config\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430737 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e642295d-af6e-4208-b077-a8b85d3047fb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.430872 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsc7s\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-kube-api-access-rsc7s\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-registry-certificates\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431104 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8jks\" (UniqueName: \"kubernetes.io/projected/53637ef2-e0c4-4e94-aca6-44b2d4bfc500-kube-api-access-b8jks\") pod \"downloads-7954f5f757-dnjv7\" (UID: \"53637ef2-e0c4-4e94-aca6-44b2d4bfc500\") " pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-ca\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431514 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz95f\" (UniqueName: \"kubernetes.io/projected/77accdd7-2331-4e73-b631-e4895dc304b3-kube-api-access-lz95f\") pod \"migrator-59844c95c7-n6dhl\" (UID: \"77accdd7-2331-4e73-b631-e4895dc304b3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431551 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dce6249-5f3e-43e9-be33-576c7acee927-serving-cert\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431580 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-client\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431682 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba95982d-4bef-422c-99d9-16fa4f0160ed-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431783 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e642295d-af6e-4208-b077-a8b85d3047fb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431852 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq9fw\" (UniqueName: \"kubernetes.io/projected/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-kube-api-access-mq9fw\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431885 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-trusted-ca\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431915 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-machine-approver-tls\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.431975 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e83f0f60-c4e8-43b1-9949-4aef1eb65994-audit-dir\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432002 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56w28\" (UniqueName: \"kubernetes.io/projected/e83f0f60-c4e8-43b1-9949-4aef1eb65994-kube-api-access-56w28\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432035 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-serving-cert\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432070 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb9w5\" (UniqueName: \"kubernetes.io/projected/7a29e45e-3063-4a6d-b215-0765aa2df10a-kube-api-access-nb9w5\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432102 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-default-certificate\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432127 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432169 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e642295d-af6e-4208-b077-a8b85d3047fb-config\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432205 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cab47b19-b54c-4498-90cf-2281376ffa72-tmpfs\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432250 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e83f0f60-c4e8-43b1-9949-4aef1eb65994-node-pullsecrets\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432281 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/36216347-d1f5-4db3-8bb8-6654caba4839-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432310 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceefca77-376e-4c60-8445-516d1d197f77-serving-cert\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-srv-cert\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432393 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cab47b19-b54c-4498-90cf-2281376ffa72-apiservice-cert\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432462 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-profile-collector-cert\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432520 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-service-ca\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432612 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfdjg\" (UniqueName: \"kubernetes.io/projected/2dce6249-5f3e-43e9-be33-576c7acee927-kube-api-access-vfdjg\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-client-ca\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqg62\" (UniqueName: \"kubernetes.io/projected/ba95982d-4bef-422c-99d9-16fa4f0160ed-kube-api-access-kqg62\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.432894 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433019 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-audit\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433228 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/36216347-d1f5-4db3-8bb8-6654caba4839-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433267 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-config\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433371 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chr8c\" (UniqueName: \"kubernetes.io/projected/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-kube-api-access-chr8c\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433408 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-config\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433482 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdkht\" (UniqueName: \"kubernetes.io/projected/a31db46c-756f-4dff-bf2a-24c900683bb8-kube-api-access-hdkht\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/545a5916-91fb-4243-b825-fc09e352c3d5-metrics-tls\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433657 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-registry-tls\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433727 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-service-ca-bundle\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433784 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-serving-cert\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433804 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-image-import-ca\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433826 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-metrics-certs\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433848 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-bound-sa-token\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433891 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b3716ff0-7157-4406-be92-89893b24a9ef-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-68sgp\" (UID: \"b3716ff0-7157-4406-be92-89893b24a9ef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.433923 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xgkv\" (UniqueName: \"kubernetes.io/projected/ceefca77-376e-4c60-8445-516d1d197f77-kube-api-access-5xgkv\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.434060 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-etcd-client\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.434174 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/146577d7-d2c0-4bf4-9d6b-236967454f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xnjf9\" (UID: \"146577d7-d2c0-4bf4-9d6b-236967454f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.434236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.434559 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:45.934544397 +0000 UTC m=+142.618500824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.435175 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1200b8de-e5c0-4315-a8bc-0284c5963a76-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-qdks2\" (UID: \"1200b8de-e5c0-4315-a8bc-0284c5963a76\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535312 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-profile-collector-cert\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535585 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-client-ca\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535605 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-service-ca\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535619 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfdjg\" (UniqueName: \"kubernetes.io/projected/2dce6249-5f3e-43e9-be33-576c7acee927-kube-api-access-vfdjg\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535639 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e308a4a1-6626-42c8-9e3b-8823beb2f204-config-volume\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535679 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-audit\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535698 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfhmv\" (UniqueName: \"kubernetes.io/projected/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-kube-api-access-xfhmv\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535716 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqg62\" (UniqueName: \"kubernetes.io/projected/ba95982d-4bef-422c-99d9-16fa4f0160ed-kube-api-access-kqg62\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535748 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bxc4\" (UniqueName: \"kubernetes.io/projected/577e54e1-5224-444f-823f-91d4d9f816c1-kube-api-access-7bxc4\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535765 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-mountpoint-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535785 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhtpf\" (UniqueName: \"kubernetes.io/projected/7e29913e-123d-4932-a84c-d9f3457fde27-kube-api-access-lhtpf\") pod \"package-server-manager-789f6589d5-jsgtd\" (UID: \"7e29913e-123d-4932-a84c-d9f3457fde27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535801 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-config\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535819 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-registration-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535847 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/36216347-d1f5-4db3-8bb8-6654caba4839-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535865 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chr8c\" (UniqueName: \"kubernetes.io/projected/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-kube-api-access-chr8c\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535883 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-config\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535900 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2360c01d-f5a8-4538-9d48-1ba58afede08-certs\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.535989 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdkht\" (UniqueName: \"kubernetes.io/projected/a31db46c-756f-4dff-bf2a-24c900683bb8-kube-api-access-hdkht\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-srv-cert\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536038 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/545a5916-91fb-4243-b825-fc09e352c3d5-metrics-tls\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536054 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-registry-tls\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536072 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqgmh\" (UniqueName: \"kubernetes.io/projected/f96684e9-9786-4245-a275-52a6da7c923a-kube-api-access-kqgmh\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-service-ca-bundle\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536124 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-serving-cert\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536144 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-serving-cert\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536161 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536216 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-image-import-ca\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536237 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536258 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-metrics-certs\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536275 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xgkv\" (UniqueName: \"kubernetes.io/projected/ceefca77-376e-4c60-8445-516d1d197f77-kube-api-access-5xgkv\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536332 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536355 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-bound-sa-token\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b3716ff0-7157-4406-be92-89893b24a9ef-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-68sgp\" (UID: \"b3716ff0-7157-4406-be92-89893b24a9ef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536396 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/146577d7-d2c0-4bf4-9d6b-236967454f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xnjf9\" (UID: \"146577d7-d2c0-4bf4-9d6b-236967454f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536419 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltjn7\" (UniqueName: \"kubernetes.io/projected/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-kube-api-access-ltjn7\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536437 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536459 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536486 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-etcd-client\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536513 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2360c01d-f5a8-4538-9d48-1ba58afede08-node-bootstrap-token\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536538 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536562 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536581 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cab47b19-b54c-4498-90cf-2281376ffa72-webhook-cert\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536596 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-etcd-serving-ca\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536626 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a31db46c-756f-4dff-bf2a-24c900683bb8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536642 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536669 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba95982d-4bef-422c-99d9-16fa4f0160ed-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536688 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-stats-auth\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536706 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c75d505-4bd3-45be-a65a-e532e517449d-serving-cert\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536722 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536743 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f96684e9-9786-4245-a275-52a6da7c923a-signing-key\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536764 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58nbl\" (UniqueName: \"kubernetes.io/projected/e308a4a1-6626-42c8-9e3b-8823beb2f204-kube-api-access-58nbl\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536785 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545a5916-91fb-4243-b825-fc09e352c3d5-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536808 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-auth-proxy-config\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536823 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kt6lg\" (UniqueName: \"kubernetes.io/projected/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-kube-api-access-kt6lg\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hvz4\" (UniqueName: \"kubernetes.io/projected/b3716ff0-7157-4406-be92-89893b24a9ef-kube-api-access-8hvz4\") pod \"multus-admission-controller-857f4d67dd-68sgp\" (UID: \"b3716ff0-7157-4406-be92-89893b24a9ef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536855 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536872 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdlmb\" (UniqueName: \"kubernetes.io/projected/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-kube-api-access-xdlmb\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536889 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnrsk\" (UniqueName: \"kubernetes.io/projected/cab47b19-b54c-4498-90cf-2281376ffa72-kube-api-access-bnrsk\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536906 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0c75d505-4bd3-45be-a65a-e532e517449d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536923 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28beec1e-6abb-40ef-8d38-d567a20f9dd8-cert\") pod \"ingress-canary-rvspl\" (UID: \"28beec1e-6abb-40ef-8d38-d567a20f9dd8\") " pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536962 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmj57\" (UniqueName: \"kubernetes.io/projected/dd0f7444-df82-40cd-aaa3-60250afa1f36-kube-api-access-gmj57\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536982 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-config\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.536998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545a5916-91fb-4243-b825-fc09e352c3d5-trusted-ca\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537013 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537079 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a31db46c-756f-4dff-bf2a-24c900683bb8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537121 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-config\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537166 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/577e54e1-5224-444f-823f-91d4d9f816c1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a29e45e-3063-4a6d-b215-0765aa2df10a-service-ca-bundle\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537202 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537234 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537256 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wvpn\" (UniqueName: \"kubernetes.io/projected/146577d7-d2c0-4bf4-9d6b-236967454f83-kube-api-access-4wvpn\") pod \"control-plane-machine-set-operator-78cbb6b69f-xnjf9\" (UID: \"146577d7-d2c0-4bf4-9d6b-236967454f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537278 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-config\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537311 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nvql\" (UniqueName: \"kubernetes.io/projected/545a5916-91fb-4243-b825-fc09e352c3d5-kube-api-access-5nvql\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537331 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537346 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e308a4a1-6626-42c8-9e3b-8823beb2f204-metrics-tls\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537363 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-encryption-config\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537380 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f96684e9-9786-4245-a275-52a6da7c923a-signing-cabundle\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537395 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/577e54e1-5224-444f-823f-91d4d9f816c1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537411 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e642295d-af6e-4208-b077-a8b85d3047fb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537428 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-dir\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537443 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwcrf\" (UniqueName: \"kubernetes.io/projected/0c75d505-4bd3-45be-a65a-e532e517449d-kube-api-access-jwcrf\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537458 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8x7s\" (UniqueName: \"kubernetes.io/projected/7bb6de6b-100d-4acf-8d51-d372948383ff-kube-api-access-v8x7s\") pod \"dns-operator-744455d44c-6xzbk\" (UID: \"7bb6de6b-100d-4acf-8d51-d372948383ff\") " pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537483 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsc7s\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-kube-api-access-rsc7s\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537522 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-registry-certificates\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537542 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8jks\" (UniqueName: \"kubernetes.io/projected/53637ef2-e0c4-4e94-aca6-44b2d4bfc500-kube-api-access-b8jks\") pod \"downloads-7954f5f757-dnjv7\" (UID: \"53637ef2-e0c4-4e94-aca6-44b2d4bfc500\") " pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537569 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-ca\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537586 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgch2\" (UniqueName: \"kubernetes.io/projected/28beec1e-6abb-40ef-8d38-d567a20f9dd8-kube-api-access-tgch2\") pod \"ingress-canary-rvspl\" (UID: \"28beec1e-6abb-40ef-8d38-d567a20f9dd8\") " pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537601 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bb6de6b-100d-4acf-8d51-d372948383ff-metrics-tls\") pod \"dns-operator-744455d44c-6xzbk\" (UID: \"7bb6de6b-100d-4acf-8d51-d372948383ff\") " pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537618 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7e29913e-123d-4932-a84c-d9f3457fde27-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jsgtd\" (UID: \"7e29913e-123d-4932-a84c-d9f3457fde27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537635 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537651 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-socket-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537668 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dce6249-5f3e-43e9-be33-576c7acee927-serving-cert\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-client\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537702 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2l6cs\" (UniqueName: \"kubernetes.io/projected/2ac8002f-f652-4341-93f7-ad4f828b564d-kube-api-access-2l6cs\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537719 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz95f\" (UniqueName: \"kubernetes.io/projected/77accdd7-2331-4e73-b631-e4895dc304b3-kube-api-access-lz95f\") pod \"migrator-59844c95c7-n6dhl\" (UID: \"77accdd7-2331-4e73-b631-e4895dc304b3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e642295d-af6e-4208-b077-a8b85d3047fb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537765 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537841 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-csi-data-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537878 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba95982d-4bef-422c-99d9-16fa4f0160ed-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537903 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-config\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537923 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd0f7444-df82-40cd-aaa3-60250afa1f36-config-volume\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.537988 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-policies\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538006 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538027 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq9fw\" (UniqueName: \"kubernetes.io/projected/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-kube-api-access-mq9fw\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538043 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnpvg\" (UniqueName: \"kubernetes.io/projected/2360c01d-f5a8-4538-9d48-1ba58afede08-kube-api-access-nnpvg\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538062 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-trusted-ca\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538078 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-machine-approver-tls\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e83f0f60-c4e8-43b1-9949-4aef1eb65994-audit-dir\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538112 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56w28\" (UniqueName: \"kubernetes.io/projected/e83f0f60-c4e8-43b1-9949-4aef1eb65994-kube-api-access-56w28\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538129 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-serving-cert\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538147 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb9w5\" (UniqueName: \"kubernetes.io/projected/7a29e45e-3063-4a6d-b215-0765aa2df10a-kube-api-access-nb9w5\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538165 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538191 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-default-certificate\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538206 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-plugins-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538223 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd0f7444-df82-40cd-aaa3-60250afa1f36-secret-volume\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538241 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cab47b19-b54c-4498-90cf-2281376ffa72-tmpfs\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538592 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8668\" (UniqueName: \"kubernetes.io/projected/7bc1de70-8871-4225-8a3e-d6f849fc864a-kube-api-access-c8668\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-925n5\" (UniqueName: \"kubernetes.io/projected/121f5347-8bf8-4d6c-ab81-3f22c87275bc-kube-api-access-925n5\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e642295d-af6e-4208-b077-a8b85d3047fb-config\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538672 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e83f0f60-c4e8-43b1-9949-4aef1eb65994-node-pullsecrets\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538694 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/36216347-d1f5-4db3-8bb8-6654caba4839-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceefca77-376e-4c60-8445-516d1d197f77-serving-cert\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538744 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cab47b19-b54c-4498-90cf-2281376ffa72-apiservice-cert\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.538770 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-srv-cert\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.541203 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.04116826 +0000 UTC m=+142.725124687 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.542732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-client-ca\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.542919 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-profile-collector-cert\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.543954 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/545a5916-91fb-4243-b825-fc09e352c3d5-trusted-ca\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.544825 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-audit\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.545972 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.548520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a31db46c-756f-4dff-bf2a-24c900683bb8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.549876 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-service-ca\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.552123 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-service-ca-bundle\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.552844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-config\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.553603 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7a29e45e-3063-4a6d-b215-0765aa2df10a-service-ca-bundle\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.554460 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.557384 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.557946 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-serving-cert\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.558507 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-config\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.559788 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.559923 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-config\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.560970 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/36216347-d1f5-4db3-8bb8-6654caba4839-ca-trust-extracted\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.576451 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/b3716ff0-7157-4406-be92-89893b24a9ef-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-68sgp\" (UID: \"b3716ff0-7157-4406-be92-89893b24a9ef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.577093 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ba95982d-4bef-422c-99d9-16fa4f0160ed-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.579315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-image-import-ca\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.579688 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.579810 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/545a5916-91fb-4243-b825-fc09e352c3d5-metrics-tls\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.580622 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba95982d-4bef-422c-99d9-16fa4f0160ed-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.589008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-encryption-config\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.589047 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/e83f0f60-c4e8-43b1-9949-4aef1eb65994-node-pullsecrets\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.589071 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/e83f0f60-c4e8-43b1-9949-4aef1eb65994-etcd-client\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.589869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-ca\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.591138 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.591826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-auth-proxy-config\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.591848 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-config\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.595564 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-stats-auth\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.596892 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-config\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.597652 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/cab47b19-b54c-4498-90cf-2281376ffa72-tmpfs\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.599115 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-metrics-certs\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.599566 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e83f0f60-c4e8-43b1-9949-4aef1eb65994-audit-dir\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.603053 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.604559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.608958 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e642295d-af6e-4208-b077-a8b85d3047fb-config\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.609044 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2dce6249-5f3e-43e9-be33-576c7acee927-etcd-client\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.609271 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2dce6249-5f3e-43e9-be33-576c7acee927-serving-cert\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.610210 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-registry-certificates\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.611625 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-serving-cert\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.617091 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfdjg\" (UniqueName: \"kubernetes.io/projected/2dce6249-5f3e-43e9-be33-576c7acee927-kube-api-access-vfdjg\") pod \"etcd-operator-b45778765-6g2vd\" (UID: \"2dce6249-5f3e-43e9-be33-576c7acee927\") " pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.618493 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/e83f0f60-c4e8-43b1-9949-4aef1eb65994-etcd-serving-ca\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.620323 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cab47b19-b54c-4498-90cf-2281376ffa72-apiservice-cert\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.626830 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sw9lq"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.627479 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cab47b19-b54c-4498-90cf-2281376ffa72-webhook-cert\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.628094 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a31db46c-756f-4dff-bf2a-24c900683bb8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.629809 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.630017 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceefca77-376e-4c60-8445-516d1d197f77-serving-cert\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.630831 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7a29e45e-3063-4a6d-b215-0765aa2df10a-default-certificate\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.630850 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/146577d7-d2c0-4bf4-9d6b-236967454f83-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-xnjf9\" (UID: \"146577d7-d2c0-4bf4-9d6b-236967454f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.632753 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/36216347-d1f5-4db3-8bb8-6654caba4839-installation-pull-secrets\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.633614 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqg62\" (UniqueName: \"kubernetes.io/projected/ba95982d-4bef-422c-99d9-16fa4f0160ed-kube-api-access-kqg62\") pod \"openshift-controller-manager-operator-756b6f6bc6-6ls8q\" (UID: \"ba95982d-4bef-422c-99d9-16fa4f0160ed\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.635339 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-machine-approver-tls\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.637360 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-srv-cert\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.638343 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e642295d-af6e-4208-b077-a8b85d3047fb-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.638569 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-trusted-ca\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.641318 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-bound-sa-token\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: W1125 09:50:45.643370 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod65f3051c_91c1_44bc_a990_4f0a1165b4d2.slice/crio-5e8f0d0f74785d5347ffb8cc328e514a8019818efa93e4e9c00c1cf06301044b WatchSource:0}: Error finding container 5e8f0d0f74785d5347ffb8cc328e514a8019818efa93e4e9c00c1cf06301044b: Status 404 returned error can't find the container with id 5e8f0d0f74785d5347ffb8cc328e514a8019818efa93e4e9c00c1cf06301044b Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.646210 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-registry-tls\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648155 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-serving-cert\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648181 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648202 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648266 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltjn7\" (UniqueName: \"kubernetes.io/projected/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-kube-api-access-ltjn7\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648285 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648308 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648330 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2360c01d-f5a8-4538-9d48-1ba58afede08-node-bootstrap-token\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648359 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648375 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648417 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c75d505-4bd3-45be-a65a-e532e517449d-serving-cert\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648433 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648448 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f96684e9-9786-4245-a275-52a6da7c923a-signing-key\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648484 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58nbl\" (UniqueName: \"kubernetes.io/projected/e308a4a1-6626-42c8-9e3b-8823beb2f204-kube-api-access-58nbl\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648521 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648553 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0c75d505-4bd3-45be-a65a-e532e517449d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28beec1e-6abb-40ef-8d38-d567a20f9dd8-cert\") pod \"ingress-canary-rvspl\" (UID: \"28beec1e-6abb-40ef-8d38-d567a20f9dd8\") " pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648602 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmj57\" (UniqueName: \"kubernetes.io/projected/dd0f7444-df82-40cd-aaa3-60250afa1f36-kube-api-access-gmj57\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648630 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/577e54e1-5224-444f-823f-91d4d9f816c1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648691 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdkht\" (UniqueName: \"kubernetes.io/projected/a31db46c-756f-4dff-bf2a-24c900683bb8-kube-api-access-hdkht\") pod \"openshift-apiserver-operator-796bbdcf4f-lpbhm\" (UID: \"a31db46c-756f-4dff-bf2a-24c900683bb8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.650009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.650639 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.648709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e308a4a1-6626-42c8-9e3b-8823beb2f204-metrics-tls\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.653699 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.654984 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/0c75d505-4bd3-45be-a65a-e532e517449d-available-featuregates\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.655130 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.655420 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-dir\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.655539 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.155525302 +0000 UTC m=+142.839481729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.655627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f96684e9-9786-4245-a275-52a6da7c923a-signing-cabundle\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.656200 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.656284 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0c75d505-4bd3-45be-a65a-e532e517449d-serving-cert\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.656369 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/577e54e1-5224-444f-823f-91d4d9f816c1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.656534 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz95f\" (UniqueName: \"kubernetes.io/projected/77accdd7-2331-4e73-b631-e4895dc304b3-kube-api-access-lz95f\") pod \"migrator-59844c95c7-n6dhl\" (UID: \"77accdd7-2331-4e73-b631-e4895dc304b3\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657069 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f96684e9-9786-4245-a275-52a6da7c923a-signing-cabundle\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657101 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-dir\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657226 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657363 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwcrf\" (UniqueName: \"kubernetes.io/projected/0c75d505-4bd3-45be-a65a-e532e517449d-kube-api-access-jwcrf\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657402 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8x7s\" (UniqueName: \"kubernetes.io/projected/7bb6de6b-100d-4acf-8d51-d372948383ff-kube-api-access-v8x7s\") pod \"dns-operator-744455d44c-6xzbk\" (UID: \"7bb6de6b-100d-4acf-8d51-d372948383ff\") " pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657592 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgch2\" (UniqueName: \"kubernetes.io/projected/28beec1e-6abb-40ef-8d38-d567a20f9dd8-kube-api-access-tgch2\") pod \"ingress-canary-rvspl\" (UID: \"28beec1e-6abb-40ef-8d38-d567a20f9dd8\") " pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657611 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bb6de6b-100d-4acf-8d51-d372948383ff-metrics-tls\") pod \"dns-operator-744455d44c-6xzbk\" (UID: \"7bb6de6b-100d-4acf-8d51-d372948383ff\") " pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657654 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7e29913e-123d-4932-a84c-d9f3457fde27-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jsgtd\" (UID: \"7e29913e-123d-4932-a84c-d9f3457fde27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657683 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-socket-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657732 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2l6cs\" (UniqueName: \"kubernetes.io/projected/2ac8002f-f652-4341-93f7-ad4f828b564d-kube-api-access-2l6cs\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657764 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657783 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-csi-data-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657816 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-config\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657837 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd0f7444-df82-40cd-aaa3-60250afa1f36-config-volume\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657876 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-policies\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657899 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657931 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnpvg\" (UniqueName: \"kubernetes.io/projected/2360c01d-f5a8-4538-9d48-1ba58afede08-kube-api-access-nnpvg\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658008 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-plugins-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658029 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd0f7444-df82-40cd-aaa3-60250afa1f36-secret-volume\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658050 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8668\" (UniqueName: \"kubernetes.io/projected/7bc1de70-8871-4225-8a3e-d6f849fc864a-kube-api-access-c8668\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658067 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-925n5\" (UniqueName: \"kubernetes.io/projected/121f5347-8bf8-4d6c-ab81-3f22c87275bc-kube-api-access-925n5\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658114 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e308a4a1-6626-42c8-9e3b-8823beb2f204-config-volume\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658138 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfhmv\" (UniqueName: \"kubernetes.io/projected/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-kube-api-access-xfhmv\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658163 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bxc4\" (UniqueName: \"kubernetes.io/projected/577e54e1-5224-444f-823f-91d4d9f816c1-kube-api-access-7bxc4\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658196 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhtpf\" (UniqueName: \"kubernetes.io/projected/7e29913e-123d-4932-a84c-d9f3457fde27-kube-api-access-lhtpf\") pod \"package-server-manager-789f6589d5-jsgtd\" (UID: \"7e29913e-123d-4932-a84c-d9f3457fde27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658221 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-mountpoint-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658261 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-registration-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658288 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2360c01d-f5a8-4538-9d48-1ba58afede08-certs\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658307 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658328 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-srv-cert\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658349 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqgmh\" (UniqueName: \"kubernetes.io/projected/f96684e9-9786-4245-a275-52a6da7c923a-kube-api-access-kqgmh\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658426 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/577e54e1-5224-444f-823f-91d4d9f816c1-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658478 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658544 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-mountpoint-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658700 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-plugins-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.658830 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-registration-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.657618 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/577e54e1-5224-444f-823f-91d4d9f816c1-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.659171 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-profile-collector-cert\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.659439 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.659452 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f96684e9-9786-4245-a275-52a6da7c923a-signing-key\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.659525 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-serving-cert\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.659599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-csi-data-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.659914 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/2360c01d-f5a8-4538-9d48-1ba58afede08-node-bootstrap-token\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.660155 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.660216 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2ac8002f-f652-4341-93f7-ad4f828b564d-socket-dir\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.660844 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-policies\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.661349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e308a4a1-6626-42c8-9e3b-8823beb2f204-config-volume\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.661748 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-config\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.662068 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.662846 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/e308a4a1-6626-42c8-9e3b-8823beb2f204-metrics-tls\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.663016 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7bb6de6b-100d-4acf-8d51-d372948383ff-metrics-tls\") pod \"dns-operator-744455d44c-6xzbk\" (UID: \"7bb6de6b-100d-4acf-8d51-d372948383ff\") " pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.663095 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd0f7444-df82-40cd-aaa3-60250afa1f36-secret-volume\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.666860 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-srv-cert\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.667144 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28beec1e-6abb-40ef-8d38-d567a20f9dd8-cert\") pod \"ingress-canary-rvspl\" (UID: \"28beec1e-6abb-40ef-8d38-d567a20f9dd8\") " pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.667450 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.667609 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.670348 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd0f7444-df82-40cd-aaa3-60250afa1f36-config-volume\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.671629 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/2360c01d-f5a8-4538-9d48-1ba58afede08-certs\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.671819 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7e29913e-123d-4932-a84c-d9f3457fde27-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jsgtd\" (UID: \"7e29913e-123d-4932-a84c-d9f3457fde27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.679249 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-6djth"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.687400 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.693430 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e642295d-af6e-4208-b077-a8b85d3047fb-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-fpbgc\" (UID: \"e642295d-af6e-4208-b077-a8b85d3047fb\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.717260 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/54f903ee-f4e2-44f5-a10c-ee204f6eaf62-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rh2ph\" (UID: \"54f903ee-f4e2-44f5-a10c-ee204f6eaf62\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.732857 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hfbwq" event={"ID":"60183186-5dd7-4632-96ef-2e87fb0c821a","Type":"ContainerStarted","Data":"5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845"} Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.733548 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wvpn\" (UniqueName: \"kubernetes.io/projected/146577d7-d2c0-4bf4-9d6b-236967454f83-kube-api-access-4wvpn\") pod \"control-plane-machine-set-operator-78cbb6b69f-xnjf9\" (UID: \"146577d7-d2c0-4bf4-9d6b-236967454f83\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.733585 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hfbwq" event={"ID":"60183186-5dd7-4632-96ef-2e87fb0c821a","Type":"ContainerStarted","Data":"5079f08021b1945fcd94cfdf67fd311c165be4bcb058dc23aed30c3a1da78604"} Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.733865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6djth" event={"ID":"c67403f1-0cf5-4b25-93dd-36253a6135e2","Type":"ContainerStarted","Data":"49ae26a4bb5f2d4ae6e0603cd19d92b7c4902544eaa49183c3eb6103aba54e9a"} Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.734711 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" event={"ID":"65f3051c-91c1-44bc-a990-4f0a1165b4d2","Type":"ContainerStarted","Data":"5e8f0d0f74785d5347ffb8cc328e514a8019818efa93e4e9c00c1cf06301044b"} Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.735528 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" event={"ID":"1413207c-fe4e-4f65-8d8f-605ba4474b63","Type":"ContainerStarted","Data":"1ec9d39cb8565d732ccabf8e33c9cf7991f50963b0e5c7034a852e0f1da5ac41"} Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.759360 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.759553 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.259523219 +0000 UTC m=+142.943479646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.759852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.760780 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.260769295 +0000 UTC m=+142.944725922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.765552 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nvql\" (UniqueName: \"kubernetes.io/projected/545a5916-91fb-4243-b825-fc09e352c3d5-kube-api-access-5nvql\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.777914 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.780692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chr8c\" (UniqueName: \"kubernetes.io/projected/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-kube-api-access-chr8c\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.808332 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xgkv\" (UniqueName: \"kubernetes.io/projected/ceefca77-376e-4c60-8445-516d1d197f77-kube-api-access-5xgkv\") pod \"route-controller-manager-6576b87f9c-dw7jd\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.821123 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsc7s\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-kube-api-access-rsc7s\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.826838 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.838257 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.843067 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kt6lg\" (UniqueName: \"kubernetes.io/projected/4ee74164-5b7d-4ece-8741-e5a51fe5fcc7-kube-api-access-kt6lg\") pod \"machine-approver-56656f9798-kdppf\" (UID: \"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.849107 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.853538 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.860899 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.862116 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.362095897 +0000 UTC m=+143.046052324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.870849 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq9fw\" (UniqueName: \"kubernetes.io/projected/b2a3d4de-2095-4c4e-a490-6fc9f98dafa4-kube-api-access-mq9fw\") pod \"authentication-operator-69f744f599-v7qs6\" (UID: \"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.881754 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8jks\" (UniqueName: \"kubernetes.io/projected/53637ef2-e0c4-4e94-aca6-44b2d4bfc500-kube-api-access-b8jks\") pod \"downloads-7954f5f757-dnjv7\" (UID: \"53637ef2-e0c4-4e94-aca6-44b2d4bfc500\") " pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.892394 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.893916 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.900034 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hvz4\" (UniqueName: \"kubernetes.io/projected/b3716ff0-7157-4406-be92-89893b24a9ef-kube-api-access-8hvz4\") pod \"multus-admission-controller-857f4d67dd-68sgp\" (UID: \"b3716ff0-7157-4406-be92-89893b24a9ef\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.902719 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.921242 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdlmb\" (UniqueName: \"kubernetes.io/projected/98f1f414-9fa6-45ba-a9cc-33f8b191abf2-kube-api-access-xdlmb\") pod \"catalog-operator-68c6474976-hbzzc\" (UID: \"98f1f414-9fa6-45ba-a9cc-33f8b191abf2\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.923742 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.936100 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.944855 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnrsk\" (UniqueName: \"kubernetes.io/projected/cab47b19-b54c-4498-90cf-2281376ffa72-kube-api-access-bnrsk\") pod \"packageserver-d55dfcdfc-2xfmf\" (UID: \"cab47b19-b54c-4498-90cf-2281376ffa72\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.953725 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.956862 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.957473 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/545a5916-91fb-4243-b825-fc09e352c3d5-bound-sa-token\") pod \"ingress-operator-5b745b69d9-sxpvs\" (UID: \"545a5916-91fb-4243-b825-fc09e352c3d5\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.957721 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.961901 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-9pvq7"] Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.962818 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:45 crc kubenswrapper[4926]: E1125 09:50:45.963198 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.463184482 +0000 UTC m=+143.147140899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.972643 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56w28\" (UniqueName: \"kubernetes.io/projected/e83f0f60-c4e8-43b1-9949-4aef1eb65994-kube-api-access-56w28\") pod \"apiserver-76f77b778f-bd494\" (UID: \"e83f0f60-c4e8-43b1-9949-4aef1eb65994\") " pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:45 crc kubenswrapper[4926]: I1125 09:50:45.994514 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb9w5\" (UniqueName: \"kubernetes.io/projected/7a29e45e-3063-4a6d-b215-0765aa2df10a-kube-api-access-nb9w5\") pod \"router-default-5444994796-bngtn\" (UID: \"7a29e45e-3063-4a6d-b215-0765aa2df10a\") " pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.029719 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/e3eb25a8-17ed-45b5-a875-42b9d33c54e7-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-srlnm\" (UID: \"e3eb25a8-17ed-45b5-a875-42b9d33c54e7\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.030082 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.038806 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.042898 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltjn7\" (UniqueName: \"kubernetes.io/projected/0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2-kube-api-access-ltjn7\") pod \"service-ca-operator-777779d784-dvtjd\" (UID: \"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.054882 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.055445 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.055491 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.060271 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58nbl\" (UniqueName: \"kubernetes.io/projected/e308a4a1-6626-42c8-9e3b-8823beb2f204-kube-api-access-58nbl\") pod \"dns-default-stkz9\" (UID: \"e308a4a1-6626-42c8-9e3b-8823beb2f204\") " pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.063716 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.064140 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.564120813 +0000 UTC m=+143.248077240 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.087275 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmj57\" (UniqueName: \"kubernetes.io/projected/dd0f7444-df82-40cd-aaa3-60250afa1f36-kube-api-access-gmj57\") pod \"collect-profiles-29401065-qbgkq\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.093322 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.115889 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwcrf\" (UniqueName: \"kubernetes.io/projected/0c75d505-4bd3-45be-a65a-e532e517449d-kube-api-access-jwcrf\") pod \"openshift-config-operator-7777fb866f-kjdgl\" (UID: \"0c75d505-4bd3-45be-a65a-e532e517449d\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.120766 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8x7s\" (UniqueName: \"kubernetes.io/projected/7bb6de6b-100d-4acf-8d51-d372948383ff-kube-api-access-v8x7s\") pod \"dns-operator-744455d44c-6xzbk\" (UID: \"7bb6de6b-100d-4acf-8d51-d372948383ff\") " pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.143406 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.153722 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgch2\" (UniqueName: \"kubernetes.io/projected/28beec1e-6abb-40ef-8d38-d567a20f9dd8-kube-api-access-tgch2\") pod \"ingress-canary-rvspl\" (UID: \"28beec1e-6abb-40ef-8d38-d567a20f9dd8\") " pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.158454 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqgmh\" (UniqueName: \"kubernetes.io/projected/f96684e9-9786-4245-a275-52a6da7c923a-kube-api-access-kqgmh\") pod \"service-ca-9c57cc56f-mztxj\" (UID: \"f96684e9-9786-4245-a275-52a6da7c923a\") " pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.161258 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.164743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.165078 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.665062694 +0000 UTC m=+143.349019121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.168081 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.178626 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bxc4\" (UniqueName: \"kubernetes.io/projected/577e54e1-5224-444f-823f-91d4d9f816c1-kube-api-access-7bxc4\") pod \"kube-storage-version-migrator-operator-b67b599dd-jq954\" (UID: \"577e54e1-5224-444f-823f-91d4d9f816c1\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.200952 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhtpf\" (UniqueName: \"kubernetes.io/projected/7e29913e-123d-4932-a84c-d9f3457fde27-kube-api-access-lhtpf\") pod \"package-server-manager-789f6589d5-jsgtd\" (UID: \"7e29913e-123d-4932-a84c-d9f3457fde27\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.203350 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.219582 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-925n5\" (UniqueName: \"kubernetes.io/projected/121f5347-8bf8-4d6c-ab81-3f22c87275bc-kube-api-access-925n5\") pod \"marketplace-operator-79b997595-nn4ms\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.230374 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.241123 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8668\" (UniqueName: \"kubernetes.io/projected/7bc1de70-8871-4225-8a3e-d6f849fc864a-kube-api-access-c8668\") pod \"oauth-openshift-558db77b4-ht99z\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.244020 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.254294 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2l6cs\" (UniqueName: \"kubernetes.io/projected/2ac8002f-f652-4341-93f7-ad4f828b564d-kube-api-access-2l6cs\") pod \"csi-hostpathplugin-hw4f2\" (UID: \"2ac8002f-f652-4341-93f7-ad4f828b564d\") " pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.265443 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.265680 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.765637485 +0000 UTC m=+143.449593912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.266232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.266599 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.766568441 +0000 UTC m=+143.450524868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.272807 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.281743 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.293326 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.299017 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.305983 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" Nov 25 09:50:46 crc kubenswrapper[4926]: W1125 09:50:46.310643 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7a29e45e_3063_4a6d_b215_0765aa2df10a.slice/crio-29d3131fcdf6342d5e135ed6d5abd7d292c7bc8bb5ab518a1cdca2d39d061071 WatchSource:0}: Error finding container 29d3131fcdf6342d5e135ed6d5abd7d292c7bc8bb5ab518a1cdca2d39d061071: Status 404 returned error can't find the container with id 29d3131fcdf6342d5e135ed6d5abd7d292c7bc8bb5ab518a1cdca2d39d061071 Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.311763 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnpvg\" (UniqueName: \"kubernetes.io/projected/2360c01d-f5a8-4538-9d48-1ba58afede08-kube-api-access-nnpvg\") pod \"machine-config-server-brh52\" (UID: \"2360c01d-f5a8-4538-9d48-1ba58afede08\") " pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.314101 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.316191 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfhmv\" (UniqueName: \"kubernetes.io/projected/b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c-kube-api-access-xfhmv\") pod \"olm-operator-6b444d44fb-cbhpl\" (UID: \"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.327408 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.331626 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.341990 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.366970 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.367373 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.867352488 +0000 UTC m=+143.551308915 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.368763 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.868754628 +0000 UTC m=+143.552711055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.369451 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.370291 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.370987 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.381720 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-brh52" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.395515 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.401748 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rvspl" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.471051 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.471348 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.971314185 +0000 UTC m=+143.655270612 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.478377 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.479117 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:46.979099695 +0000 UTC m=+143.663056122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.588718 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.588844 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.088815195 +0000 UTC m=+143.772771622 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.589489 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.590008 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.089984398 +0000 UTC m=+143.773940995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.591815 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-6g2vd"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.687952 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.690178 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl"] Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.691689 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.691917 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.191873236 +0000 UTC m=+143.875829673 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.693694 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.694237 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.194226063 +0000 UTC m=+143.878182500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.750000 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" event={"ID":"1200b8de-e5c0-4315-a8bc-0284c5963a76","Type":"ContainerStarted","Data":"b0a81409100add82fdde0e4f0163973da026132e9342863510497c0b425b74ab"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.750366 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" event={"ID":"1200b8de-e5c0-4315-a8bc-0284c5963a76","Type":"ContainerStarted","Data":"5184dedd89065a23d9aa9a60297a2562d649a0a81332bdad96d11c1a19cf203d"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.751895 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" event={"ID":"ceefca77-376e-4c60-8445-516d1d197f77","Type":"ContainerStarted","Data":"15f74ecb5d6ce2364a9188c728b963caf0a9c2ba56fe3d40de30adfb0587bfb4"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.760738 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bngtn" event={"ID":"7a29e45e-3063-4a6d-b215-0765aa2df10a","Type":"ContainerStarted","Data":"29d3131fcdf6342d5e135ed6d5abd7d292c7bc8bb5ab518a1cdca2d39d061071"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.762476 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" event={"ID":"65f3051c-91c1-44bc-a990-4f0a1165b4d2","Type":"ContainerStarted","Data":"2bb302c8a404f775fcb70e4691be106fb06de8cac1447c674f0fae23430cc2dc"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.762805 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.763893 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" event={"ID":"2dce6249-5f3e-43e9-be33-576c7acee927","Type":"ContainerStarted","Data":"00f4681c3f8ef261b378921cdc3ae1b7cae5da505473738275dd279b7dc9f4d3"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.765353 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" event={"ID":"a31db46c-756f-4dff-bf2a-24c900683bb8","Type":"ContainerStarted","Data":"bf9d2c05bc722d3893b63aafd54eab72bd55bffb4ef4872469625116b92581f5"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.765390 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" event={"ID":"a31db46c-756f-4dff-bf2a-24c900683bb8","Type":"ContainerStarted","Data":"dd9908ccbea3426c61a67ef3bccff373552387d3680e0263adbd2cac71f5ad2b"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.767139 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" event={"ID":"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7","Type":"ContainerStarted","Data":"d48f757255e88be4f0039311deddff5980c7b3cd5154f22fa068ae18e9c4cca4"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.767170 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" event={"ID":"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7","Type":"ContainerStarted","Data":"8ec79417d31ae8a2ff70cbf14bb846d7d4931079dd38abbfbada3f0e96e46f6a"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.767272 4926 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-sw9lq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.767316 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" podUID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.768879 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" event={"ID":"6a6a3048-e831-4418-baf5-7bca98a35cef","Type":"ContainerStarted","Data":"9398e2591a46b5cc8ad55f0f067fded571d9435311ecabd13e60fadfa7b40889"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.768957 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" event={"ID":"6a6a3048-e831-4418-baf5-7bca98a35cef","Type":"ContainerStarted","Data":"8e8eabb1c045022fa830827355cc144860f4bce35eaabec2babc3765fd8ed460"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.772238 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" event={"ID":"ba95982d-4bef-422c-99d9-16fa4f0160ed","Type":"ContainerStarted","Data":"67e25266deee406e669a87c313eab7df41d02336864e50d0497787b7dd6d47d2"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.772281 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" event={"ID":"ba95982d-4bef-422c-99d9-16fa4f0160ed","Type":"ContainerStarted","Data":"37ba2b2d5987b98a86542c3f53f1f3ec46e4a7400a64faeb319607344a07e645"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.776357 4926 generic.go:334] "Generic (PLEG): container finished" podID="1413207c-fe4e-4f65-8d8f-605ba4474b63" containerID="19733acd2c0a9ab1e8db7aba6e7cb28369d2a0ef4eca8d00613c2cfe0f1f99bc" exitCode=0 Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.776425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" event={"ID":"1413207c-fe4e-4f65-8d8f-605ba4474b63","Type":"ContainerDied","Data":"19733acd2c0a9ab1e8db7aba6e7cb28369d2a0ef4eca8d00613c2cfe0f1f99bc"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.778387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-6djth" event={"ID":"c67403f1-0cf5-4b25-93dd-36253a6135e2","Type":"ContainerStarted","Data":"004a439be35078b947606e5415a131864f86973a9ea717a15b62bc80e4686b34"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.778715 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.779602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" event={"ID":"54f903ee-f4e2-44f5-a10c-ee204f6eaf62","Type":"ContainerStarted","Data":"a28117ddec9389e7e272dcba46860562f4449d8a99f9bd5ef6a09323af8bc249"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.780501 4926 patch_prober.go:28] interesting pod/console-operator-58897d9998-6djth container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.780544 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-6djth" podUID="c67403f1-0cf5-4b25-93dd-36253a6135e2" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/readyz\": dial tcp 10.217.0.7:8443: connect: connection refused" Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.784417 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-brh52" event={"ID":"2360c01d-f5a8-4538-9d48-1ba58afede08","Type":"ContainerStarted","Data":"fad973c014089961b5dc3e6911a3e8bf32e0c25b22f3889b7709eb8903b1e313"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.790349 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" event={"ID":"12d97152-164d-4a5b-b3a0-ebfdb983d2d4","Type":"ContainerStarted","Data":"b4bf8019e7305a2f5d8867c946788ea353b5587eb77eb1c8b3331bef087068d6"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.790515 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" event={"ID":"12d97152-164d-4a5b-b3a0-ebfdb983d2d4","Type":"ContainerStarted","Data":"f55d19d32fb925670843bb0fee7b1f7ed0417eb00cda8801983328522f403885"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.793374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" event={"ID":"f2e3a659-20d7-47a9-9f7d-256a88c0e315","Type":"ContainerStarted","Data":"6f48e6b2d5ca5e8bde9abab6877542417f56e5c8c07fbf50a9e1a59daac148e1"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.793419 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" event={"ID":"f2e3a659-20d7-47a9-9f7d-256a88c0e315","Type":"ContainerStarted","Data":"d49dcf8c6909961866e59e40b2ca749061713bd75daf6f89877a2e058532f6ad"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.794637 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.794725 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.294704351 +0000 UTC m=+143.978660778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.795023 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.795550 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.295530684 +0000 UTC m=+143.979487111 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.795888 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" event={"ID":"ec3f0e93-5599-45c0-b6d0-5e16be0df94c","Type":"ContainerStarted","Data":"d87176bd66ecd68dd6298a59b8bb5617bd6d0dc05333963b6fdbd176acd5c5d7"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.796286 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" event={"ID":"ec3f0e93-5599-45c0-b6d0-5e16be0df94c","Type":"ContainerStarted","Data":"c7b3518168258a4fa79b502eb48a399a7bd445312cc6982d45871c2ce67aa8af"} Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.899150 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.899620 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.399583234 +0000 UTC m=+144.083539661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:46 crc kubenswrapper[4926]: I1125 09:50:46.903633 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:46 crc kubenswrapper[4926]: E1125 09:50:46.904731 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.404708829 +0000 UTC m=+144.088665456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.004835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.005101 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.505072874 +0000 UTC m=+144.189029311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.006502 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.007191 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.507167463 +0000 UTC m=+144.191123890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.045629 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf"] Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.109762 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.110247 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.610225904 +0000 UTC m=+144.294182331 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: W1125 09:50:47.118616 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcab47b19_b54c_4498_90cf_2281376ffa72.slice/crio-3188720e5c512d32fdbd8c063ba73640da9152ed68b88d9ed5bc5ed269267e3b WatchSource:0}: Error finding container 3188720e5c512d32fdbd8c063ba73640da9152ed68b88d9ed5bc5ed269267e3b: Status 404 returned error can't find the container with id 3188720e5c512d32fdbd8c063ba73640da9152ed68b88d9ed5bc5ed269267e3b Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.195716 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-hfbwq" podStartSLOduration=123.195691426 podStartE2EDuration="2m3.195691426s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:47.195415948 +0000 UTC m=+143.879372385" watchObservedRunningTime="2025-11-25 09:50:47.195691426 +0000 UTC m=+143.879647853" Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.218747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.219858 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.71984103 +0000 UTC m=+144.403797457 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.321308 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.321774 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.821754549 +0000 UTC m=+144.505710976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.424760 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.436755 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:47.936707767 +0000 UTC m=+144.620664194 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.521922 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-6ls8q" podStartSLOduration=123.521902512 podStartE2EDuration="2m3.521902512s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:47.519766781 +0000 UTC m=+144.203723218" watchObservedRunningTime="2025-11-25 09:50:47.521902512 +0000 UTC m=+144.205858939" Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.526645 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.526847 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.026818191 +0000 UTC m=+144.710774618 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.526905 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.527572 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.027564152 +0000 UTC m=+144.711520579 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.629923 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.630776 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.130754257 +0000 UTC m=+144.814710684 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.630952 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.631436 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.131420586 +0000 UTC m=+144.815377013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.709912 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-6djth" podStartSLOduration=123.70988373 podStartE2EDuration="2m3.70988373s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:47.66367501 +0000 UTC m=+144.347631437" watchObservedRunningTime="2025-11-25 09:50:47.70988373 +0000 UTC m=+144.393840177" Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.727988 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9"] Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.739755 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-qdks2" podStartSLOduration=123.739733046 podStartE2EDuration="2m3.739733046s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:47.739403647 +0000 UTC m=+144.423360074" watchObservedRunningTime="2025-11-25 09:50:47.739733046 +0000 UTC m=+144.423689473" Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.745836 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.746155 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.246124137 +0000 UTC m=+144.930080564 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.746607 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.747103 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.247095825 +0000 UTC m=+144.931052252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.810703 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-68sgp"] Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.847556 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.849048 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.349013754 +0000 UTC m=+145.032970181 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.886949 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" event={"ID":"f2e3a659-20d7-47a9-9f7d-256a88c0e315","Type":"ContainerStarted","Data":"56cb794677dac857a7e0dfc7720ee9fe9b41c1777ad54106739ae5b8d545cb13"} Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.901746 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" event={"ID":"cab47b19-b54c-4498-90cf-2281376ffa72","Type":"ContainerStarted","Data":"3188720e5c512d32fdbd8c063ba73640da9152ed68b88d9ed5bc5ed269267e3b"} Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.920481 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" event={"ID":"ec3f0e93-5599-45c0-b6d0-5e16be0df94c","Type":"ContainerStarted","Data":"377f8f0b337218cbc5456eb7e90526492d5ca5c191e66475b3e2ffa285146576"} Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.936403 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd"] Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.937596 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" podStartSLOduration=123.937568954 podStartE2EDuration="2m3.937568954s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:47.931379628 +0000 UTC m=+144.615336055" watchObservedRunningTime="2025-11-25 09:50:47.937568954 +0000 UTC m=+144.621525381" Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.956871 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:47 crc kubenswrapper[4926]: E1125 09:50:47.957512 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.457496979 +0000 UTC m=+145.141453406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.983100 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dnjv7"] Nov 25 09:50:47 crc kubenswrapper[4926]: I1125 09:50:47.985860 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bd494"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.005372 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.016772 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-v7qs6"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.031962 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-lpbhm" podStartSLOduration=124.031914288 podStartE2EDuration="2m4.031914288s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:47.993397466 +0000 UTC m=+144.677353893" watchObservedRunningTime="2025-11-25 09:50:48.031914288 +0000 UTC m=+144.715870715" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.058135 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.059760 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.559740837 +0000 UTC m=+145.243697264 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.066071 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" event={"ID":"12d97152-164d-4a5b-b3a0-ebfdb983d2d4","Type":"ContainerStarted","Data":"e6ae8e813e3d7aad6074c953bfc8bc3bbf814e4e37b338274f04ae846e330705"} Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.067179 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nn4ms"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.067298 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-stkz9"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.067405 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.067497 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm"] Nov 25 09:50:48 crc kubenswrapper[4926]: W1125 09:50:48.069110 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod53637ef2_e0c4_4e94_aca6_44b2d4bfc500.slice/crio-f98fd78f34e2d29f51aeab590379d0508c05f8ade17f187204872bd4745c44f1 WatchSource:0}: Error finding container f98fd78f34e2d29f51aeab590379d0508c05f8ade17f187204872bd4745c44f1: Status 404 returned error can't find the container with id f98fd78f34e2d29f51aeab590379d0508c05f8ade17f187204872bd4745c44f1 Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.086086 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" event={"ID":"6a6a3048-e831-4418-baf5-7bca98a35cef","Type":"ContainerStarted","Data":"a0391fa23297cea3ea1bbc5ed7e009bc1892a64dd443ccbd5b0013fd7bd85aa0"} Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.087828 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-bngtn" event={"ID":"7a29e45e-3063-4a6d-b215-0765aa2df10a","Type":"ContainerStarted","Data":"b1aa1e43179a25d2d162532ccb5815121c135843d1dd7fa697133d55d08f15d7"} Nov 25 09:50:48 crc kubenswrapper[4926]: W1125 09:50:48.109886 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode83f0f60_c4e8_43b1_9949_4aef1eb65994.slice/crio-e140126789b5c732ef32dd5ed5a9704dc73334bef10ce07b0ad8ab97867daf4c WatchSource:0}: Error finding container e140126789b5c732ef32dd5ed5a9704dc73334bef10ce07b0ad8ab97867daf4c: Status 404 returned error can't find the container with id e140126789b5c732ef32dd5ed5a9704dc73334bef10ce07b0ad8ab97867daf4c Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.137711 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" event={"ID":"77accdd7-2331-4e73-b631-e4895dc304b3","Type":"ContainerStarted","Data":"4146f201c806d064c52c0282750227b45b35c0803709bdd53de91ebd2bb7b60c"} Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.190145 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.191685 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.691669586 +0000 UTC m=+145.375626013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.230227 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" event={"ID":"e642295d-af6e-4208-b077-a8b85d3047fb","Type":"ContainerStarted","Data":"06791cd6158e8eba915cd38ec12ccf2ee88519ed35654590fed6b7e82ba4d7ae"} Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.245102 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.249510 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" event={"ID":"146577d7-d2c0-4bf4-9d6b-236967454f83","Type":"ContainerStarted","Data":"c3df02940bcaa814276b08d4b472bd0831b373b8c657a70756d3bedbb180bb02"} Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.267803 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:48 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:48 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:48 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.268133 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.285493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" event={"ID":"4ee74164-5b7d-4ece-8741-e5a51fe5fcc7","Type":"ContainerStarted","Data":"ec4a88158712af0b8e53e7f4016a033a6412b0f8c9ba958231955b8ef4b662db"} Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.316286 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.316779 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-bngtn" podStartSLOduration=124.316757732 podStartE2EDuration="2m4.316757732s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:48.287908594 +0000 UTC m=+144.971865021" watchObservedRunningTime="2025-11-25 09:50:48.316757732 +0000 UTC m=+145.000714149" Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.317332 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.817311777 +0000 UTC m=+145.501268204 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.333504 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.334864 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.336037 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-mztxj"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.336151 4926 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-dw7jd container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.336310 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" podUID="ceefca77-376e-4c60-8445-516d1d197f77" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.13:8443/healthz\": dial tcp 10.217.0.13:8443: connect: connection refused" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.351563 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.368500 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-6djth" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.413970 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-vcbkn" podStartSLOduration=124.413918405 podStartE2EDuration="2m4.413918405s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:48.358146695 +0000 UTC m=+145.042103122" watchObservedRunningTime="2025-11-25 09:50:48.413918405 +0000 UTC m=+145.097874832" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.421100 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.423949 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:48.923917159 +0000 UTC m=+145.607873586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.523112 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.523299 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.023273495 +0000 UTC m=+145.707229922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.523521 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.523879 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.023869892 +0000 UTC m=+145.707826319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.537696 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.544646 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht99z"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.546598 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rvspl"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.550040 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-ld9cl" podStartSLOduration=124.550019453 podStartE2EDuration="2m4.550019453s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:48.548423358 +0000 UTC m=+145.232379785" watchObservedRunningTime="2025-11-25 09:50:48.550019453 +0000 UTC m=+145.233975880" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.566817 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hw4f2"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.576897 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.624664 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.625054 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.12503697 +0000 UTC m=+145.808993397 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.659872 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.682503 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-9pvq7" podStartSLOduration=124.682486268 podStartE2EDuration="2m4.682486268s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:48.675478219 +0000 UTC m=+145.359434636" watchObservedRunningTime="2025-11-25 09:50:48.682486268 +0000 UTC m=+145.366442695" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.707172 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6xzbk"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.727910 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.728450 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.22843408 +0000 UTC m=+145.912390507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.774251 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd"] Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.802107 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-tqcdd" podStartSLOduration=124.802081098 podStartE2EDuration="2m4.802081098s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:48.80111541 +0000 UTC m=+145.485071837" watchObservedRunningTime="2025-11-25 09:50:48.802081098 +0000 UTC m=+145.486037525" Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.855595 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.855972 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.355951805 +0000 UTC m=+146.039908232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:48 crc kubenswrapper[4926]: I1125 09:50:48.956709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:48 crc kubenswrapper[4926]: E1125 09:50:48.957127 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.457114372 +0000 UTC m=+146.141070799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.059152 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.059525 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.559483574 +0000 UTC m=+146.243440011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.060019 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.064616 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.564593409 +0000 UTC m=+146.248549836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.163015 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.163331 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.663315597 +0000 UTC m=+146.347272024 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.258927 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kdppf" podStartSLOduration=125.258901506 podStartE2EDuration="2m5.258901506s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.19659823 +0000 UTC m=+145.880554657" watchObservedRunningTime="2025-11-25 09:50:49.258901506 +0000 UTC m=+145.942857933" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.259390 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:49 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:49 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:49 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.259471 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.274036 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.274426 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.774411716 +0000 UTC m=+146.458368143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.393928 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.394243 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.894221432 +0000 UTC m=+146.578177859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.394406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.394834 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.894826779 +0000 UTC m=+146.578783206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.402233 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rvspl" event={"ID":"28beec1e-6abb-40ef-8d38-d567a20f9dd8","Type":"ContainerStarted","Data":"9477c0d99a349bf58046dfefda35b0be2bf54833454432c361f9e6f720d98523"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.416198 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" podStartSLOduration=125.416174384 podStartE2EDuration="2m5.416174384s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.354230218 +0000 UTC m=+146.038186645" watchObservedRunningTime="2025-11-25 09:50:49.416174384 +0000 UTC m=+146.100130811" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.464213 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-brh52" podStartSLOduration=6.464188875 podStartE2EDuration="6.464188875s" podCreationTimestamp="2025-11-25 09:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.43579056 +0000 UTC m=+146.119746987" watchObservedRunningTime="2025-11-25 09:50:49.464188875 +0000 UTC m=+146.148145302" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.494805 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" event={"ID":"ceefca77-376e-4c60-8445-516d1d197f77","Type":"ContainerStarted","Data":"38ca9fda52a4f0fa63c3018975ef9de273de601b95686354ba814b27444807c4"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.496372 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.496817 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:49.996800659 +0000 UTC m=+146.680757086 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.546322 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" event={"ID":"577e54e1-5224-444f-823f-91d4d9f816c1","Type":"ContainerStarted","Data":"2e6babbc3497dd83481c9401495919094974d9a89202f1cdc9248bc0397c41be"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.563475 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dnjv7" event={"ID":"53637ef2-e0c4-4e94-aca6-44b2d4bfc500","Type":"ContainerStarted","Data":"332f01a285ac1a6bd5e3e72a6e5e99cf6ead5e964dc83a976ae1c09ff5989f44"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.563538 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dnjv7" event={"ID":"53637ef2-e0c4-4e94-aca6-44b2d4bfc500","Type":"ContainerStarted","Data":"f98fd78f34e2d29f51aeab590379d0508c05f8ade17f187204872bd4745c44f1"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.564369 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.565314 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.565356 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.591182 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" podStartSLOduration=125.591160694 podStartE2EDuration="2m5.591160694s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.59068721 +0000 UTC m=+146.274643637" watchObservedRunningTime="2025-11-25 09:50:49.591160694 +0000 UTC m=+146.275117121" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.599845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.600676 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.100660083 +0000 UTC m=+146.784616520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.606064 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.643162 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dnjv7" podStartSLOduration=125.643108576 podStartE2EDuration="2m5.643108576s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.635521701 +0000 UTC m=+146.319478118" watchObservedRunningTime="2025-11-25 09:50:49.643108576 +0000 UTC m=+146.327065003" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.645768 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" event={"ID":"7e29913e-123d-4932-a84c-d9f3457fde27","Type":"ContainerStarted","Data":"a8c2b3b2f699b04bb443da1429f4d6d7448f11248ed47d58ded203a38e17097c"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.682370 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" event={"ID":"e642295d-af6e-4208-b077-a8b85d3047fb","Type":"ContainerStarted","Data":"8298a62026a7cc51ee1c8ea58f424dd10ccdce214fe1f9e31fd61a0b9905a602"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.700659 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" event={"ID":"54f903ee-f4e2-44f5-a10c-ee204f6eaf62","Type":"ContainerStarted","Data":"849082b77ea33aaab28f934c5c40e443e4042d88291028ad68bb02a0526df203"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.701704 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.702156 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.202126459 +0000 UTC m=+146.886082886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.702309 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.704079 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.204069314 +0000 UTC m=+146.888025741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.728716 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" event={"ID":"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4","Type":"ContainerStarted","Data":"86c5f86280cb6f402d542a79019b3f547d8a07634cd2ae011881dd433b63ec33"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.729268 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-fpbgc" podStartSLOduration=125.729253848 podStartE2EDuration="2m5.729253848s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.72860037 +0000 UTC m=+146.412556797" watchObservedRunningTime="2025-11-25 09:50:49.729253848 +0000 UTC m=+146.413210275" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.769817 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" event={"ID":"146577d7-d2c0-4bf4-9d6b-236967454f83","Type":"ContainerStarted","Data":"0f6eeb30251fdfb3468573a397044da804ff16723680daea7fd539d3d03c1d3d"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.797306 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" event={"ID":"e3eb25a8-17ed-45b5-a875-42b9d33c54e7","Type":"ContainerStarted","Data":"1b4276a2e6aee7efc3c53544ca645635a37e084f1ed697b62f87057706274d7a"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.804234 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.805419 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.305402116 +0000 UTC m=+146.989358543 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.831053 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" event={"ID":"77accdd7-2331-4e73-b631-e4895dc304b3","Type":"ContainerStarted","Data":"66259f6bbb786a419c229f837da7c19770df36c1edd6a68e54cbb3c6afc936c2"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.831105 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" event={"ID":"77accdd7-2331-4e73-b631-e4895dc304b3","Type":"ContainerStarted","Data":"2a389241748a4a3cbb7af8859b0b14fda691bb3522845f19cb7292f27f016e4a"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.848569 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" podStartSLOduration=125.848536229 podStartE2EDuration="2m5.848536229s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.848087916 +0000 UTC m=+146.532044353" watchObservedRunningTime="2025-11-25 09:50:49.848536229 +0000 UTC m=+146.532492656" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.850963 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rh2ph" podStartSLOduration=125.850926057 podStartE2EDuration="2m5.850926057s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.782806606 +0000 UTC m=+146.466763033" watchObservedRunningTime="2025-11-25 09:50:49.850926057 +0000 UTC m=+146.534882484" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.859788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" event={"ID":"f96684e9-9786-4245-a275-52a6da7c923a","Type":"ContainerStarted","Data":"a6a52a87102524a37a08f743288dd783663ede8dcbadba4ad5937bf3221be718"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.859854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" event={"ID":"f96684e9-9786-4245-a275-52a6da7c923a","Type":"ContainerStarted","Data":"26aa96e16d76644e889c1af726cf5544115f4896deb1abcfde38ffbb3a599bcc"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.889522 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-xnjf9" podStartSLOduration=125.88948503 podStartE2EDuration="2m5.88948503s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.887101312 +0000 UTC m=+146.571057749" watchObservedRunningTime="2025-11-25 09:50:49.88948503 +0000 UTC m=+146.573441457" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.909550 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:49 crc kubenswrapper[4926]: E1125 09:50:49.910711 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.410697511 +0000 UTC m=+147.094653928 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.932223 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" event={"ID":"b3716ff0-7157-4406-be92-89893b24a9ef","Type":"ContainerStarted","Data":"8b9a93f461ef16b80ea9345bc3d8ddc30a4bdcdceff0d93c08f4d9e4599e4c4f"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.937192 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" event={"ID":"2dce6249-5f3e-43e9-be33-576c7acee927","Type":"ContainerStarted","Data":"48e76ef4d091a228748b3bb7a069f7de4397380f9780acd6a0e2dff823518921"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.961367 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" podStartSLOduration=125.961342837 podStartE2EDuration="2m5.961342837s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.925734847 +0000 UTC m=+146.609691274" watchObservedRunningTime="2025-11-25 09:50:49.961342837 +0000 UTC m=+146.645299264" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.961764 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-mztxj" podStartSLOduration=125.961756378 podStartE2EDuration="2m5.961756378s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:49.95968812 +0000 UTC m=+146.643644547" watchObservedRunningTime="2025-11-25 09:50:49.961756378 +0000 UTC m=+146.645712805" Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.982195 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" event={"ID":"cab47b19-b54c-4498-90cf-2281376ffa72","Type":"ContainerStarted","Data":"086218ceafcee9acce409f15d2e713258f6cff94a6eb4bb929a2847f2e9ae571"} Nov 25 09:50:49 crc kubenswrapper[4926]: I1125 09:50:49.983134 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.004600 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.010464 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.011834 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.511798857 +0000 UTC m=+147.195755284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.074762 4926 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nn4ms container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.074831 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.087833 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.087875 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" event={"ID":"121f5347-8bf8-4d6c-ab81-3f22c87275bc","Type":"ContainerStarted","Data":"1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.087897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" event={"ID":"121f5347-8bf8-4d6c-ab81-3f22c87275bc","Type":"ContainerStarted","Data":"c0f58f67bd261415d14d9c7a8937b818c5270f075cbd9d4696be93b958c4e535"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.087909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-brh52" event={"ID":"2360c01d-f5a8-4538-9d48-1ba58afede08","Type":"ContainerStarted","Data":"9e79f6653f7cf32e91ac3d0dc1953c68fe8e3bfd57e7d580b0d5ab51d7dd4a1e"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.092901 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-stkz9" event={"ID":"e308a4a1-6626-42c8-9e3b-8823beb2f204","Type":"ContainerStarted","Data":"70ec2a4bdde6ca5e9e9bb8774c41a35e7548c8f2736c183cccb9799d08838e74"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.095808 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" event={"ID":"2ac8002f-f652-4341-93f7-ad4f828b564d","Type":"ContainerStarted","Data":"16f4f39c00201f3703af381bbd6e56ffc95349ef0700651cdeecc42d240034b5"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.104220 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" event={"ID":"7bb6de6b-100d-4acf-8d51-d372948383ff","Type":"ContainerStarted","Data":"f0c9ad2aca876ca752b5f019e388897bccc3f64532921030a2135b6d3c7b960b"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.116586 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.118652 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.618638725 +0000 UTC m=+147.302595152 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.151850 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-2xfmf" podStartSLOduration=126.151825026 podStartE2EDuration="2m6.151825026s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.150540619 +0000 UTC m=+146.834497046" watchObservedRunningTime="2025-11-25 09:50:50.151825026 +0000 UTC m=+146.835781463" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.152365 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-n6dhl" podStartSLOduration=126.152357551 podStartE2EDuration="2m6.152357551s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.006180437 +0000 UTC m=+146.690136864" watchObservedRunningTime="2025-11-25 09:50:50.152357551 +0000 UTC m=+146.836313978" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.152815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" event={"ID":"0c75d505-4bd3-45be-a65a-e532e517449d","Type":"ContainerStarted","Data":"cec381df882810eae537ab3ee3b8b65e067b9bdbd94108067a50aa4291484282"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.159869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" event={"ID":"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2","Type":"ContainerStarted","Data":"b70f55d2b2a46f9c964fe6ef030e41eaa1f25f1d7181618090a250f2b2e6e684"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.160198 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" event={"ID":"0fd5c72b-5df1-4a42-a71f-7eb9d68cf2f2","Type":"ContainerStarted","Data":"1423b5b2555956127b13df606dfbc62568894fb3639b198a00b37afcfecff4d7"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.188013 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" event={"ID":"1413207c-fe4e-4f65-8d8f-605ba4474b63","Type":"ContainerStarted","Data":"ec28327db126ae936504272bc21d6105dcc7a4916eb596227dff766ff1a585f5"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.211546 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" event={"ID":"7bc1de70-8871-4225-8a3e-d6f849fc864a","Type":"ContainerStarted","Data":"a04401f0eb89dd3a9ada364e71ef03d1d267aa865b50088989ffa5087b8a2e12"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.212532 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.228441 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.228842 4926 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ht99z container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" start-of-body= Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.228996 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.230224 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.730206157 +0000 UTC m=+147.414162584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.235141 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" event={"ID":"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c","Type":"ContainerStarted","Data":"5e45e68431ceb4be689a56cb655a6108b5b21802c37664885603aef8cbb47957"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.236004 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.258135 4926 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cbhpl container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.258215 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" podUID="b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.270499 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:50 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:50 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:50 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.270851 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.271203 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" event={"ID":"dd0f7444-df82-40cd-aaa3-60250afa1f36","Type":"ContainerStarted","Data":"091b294a5025334a1f56b13080a5d8c9553d0b993299176d05642e1302748036"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.278294 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.278802 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.302275 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" event={"ID":"545a5916-91fb-4243-b825-fc09e352c3d5","Type":"ContainerStarted","Data":"917764cd8de4ffdfb0290bff2676b6f1faf3b74a850d0bc847e7fae71a7a0b74"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.325982 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-dvtjd" podStartSLOduration=126.32591616 podStartE2EDuration="2m6.32591616s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.314135806 +0000 UTC m=+146.998092233" watchObservedRunningTime="2025-11-25 09:50:50.32591616 +0000 UTC m=+147.009872587" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.327386 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-6g2vd" podStartSLOduration=126.327379232 podStartE2EDuration="2m6.327379232s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.212697801 +0000 UTC m=+146.896654228" watchObservedRunningTime="2025-11-25 09:50:50.327379232 +0000 UTC m=+147.011335659" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.331005 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.336346 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.836324575 +0000 UTC m=+147.520281202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.358787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" event={"ID":"98f1f414-9fa6-45ba-a9cc-33f8b191abf2","Type":"ContainerStarted","Data":"e3461857e8c8e199bce686ba85397e590ac9c035855c15e79c35c9218bdb57f1"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.358847 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" event={"ID":"98f1f414-9fa6-45ba-a9cc-33f8b191abf2","Type":"ContainerStarted","Data":"8463ff7343c4029a2898a35fce76f57e975a0d58400b5fc7a6fc2f7e382ca6b5"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.359981 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.364735 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" podStartSLOduration=126.36471636 podStartE2EDuration="2m6.36471636s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.364565836 +0000 UTC m=+147.048522263" watchObservedRunningTime="2025-11-25 09:50:50.36471636 +0000 UTC m=+147.048672777" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.381259 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.395546 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bd494" event={"ID":"e83f0f60-c4e8-43b1-9949-4aef1eb65994","Type":"ContainerStarted","Data":"e140126789b5c732ef32dd5ed5a9704dc73334bef10ce07b0ad8ab97867daf4c"} Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.416682 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" podStartSLOduration=126.416657552 podStartE2EDuration="2m6.416657552s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.416181899 +0000 UTC m=+147.100138326" watchObservedRunningTime="2025-11-25 09:50:50.416657552 +0000 UTC m=+147.100613979" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.440109 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.442384 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:50.942364921 +0000 UTC m=+147.626321348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.543419 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.543815 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.043796296 +0000 UTC m=+147.727752723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.556305 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.580044 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" podStartSLOduration=126.580014532 podStartE2EDuration="2m6.580014532s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.574337721 +0000 UTC m=+147.258294148" watchObservedRunningTime="2025-11-25 09:50:50.580014532 +0000 UTC m=+147.263970959" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.581357 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" podStartSLOduration=126.58134789 podStartE2EDuration="2m6.58134789s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.496123975 +0000 UTC m=+147.180080402" watchObservedRunningTime="2025-11-25 09:50:50.58134789 +0000 UTC m=+147.265304317" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.646530 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.646789 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.146772445 +0000 UTC m=+147.830728872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.656451 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" podStartSLOduration=126.656429368 podStartE2EDuration="2m6.656429368s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.647065363 +0000 UTC m=+147.331021790" watchObservedRunningTime="2025-11-25 09:50:50.656429368 +0000 UTC m=+147.340385795" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.748845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.749587 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.249570857 +0000 UTC m=+147.933527284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.800240 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" podStartSLOduration=126.800214243 podStartE2EDuration="2m6.800214243s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.798985428 +0000 UTC m=+147.482941855" watchObservedRunningTime="2025-11-25 09:50:50.800214243 +0000 UTC m=+147.484170670" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.843836 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hbzzc" podStartSLOduration=126.843814769 podStartE2EDuration="2m6.843814769s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:50.840656009 +0000 UTC m=+147.524612436" watchObservedRunningTime="2025-11-25 09:50:50.843814769 +0000 UTC m=+147.527771196" Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.849839 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.850019 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.349989704 +0000 UTC m=+148.033946131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.850081 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.850697 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.350684243 +0000 UTC m=+148.034640680 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:50 crc kubenswrapper[4926]: I1125 09:50:50.953692 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:50 crc kubenswrapper[4926]: E1125 09:50:50.954163 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.454138826 +0000 UTC m=+148.138095253 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.055928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.056383 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.556367293 +0000 UTC m=+148.240323720 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.157794 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.158051 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.658023685 +0000 UTC m=+148.341980112 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.158770 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.159165 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.659147726 +0000 UTC m=+148.343104153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.254902 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:51 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:51 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:51 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.254980 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.259379 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.259666 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.759624744 +0000 UTC m=+148.443581171 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.259760 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.260633 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.760612983 +0000 UTC m=+148.444569590 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.361114 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.361322 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.861289656 +0000 UTC m=+148.545246083 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.361556 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.361970 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.861918084 +0000 UTC m=+148.545874501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.421129 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" event={"ID":"7e29913e-123d-4932-a84c-d9f3457fde27","Type":"ContainerStarted","Data":"8fc0cb08e84cd218ee6a9e9c65a1b39128d545688b5e63aa6d09f73e0484b4a2"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.421189 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" event={"ID":"7e29913e-123d-4932-a84c-d9f3457fde27","Type":"ContainerStarted","Data":"4f45a658abb5928d0aa2c4edf75a1b04d84e76d1832dd4689a0e78b7cdceba98"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.422197 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.435500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" event={"ID":"7bb6de6b-100d-4acf-8d51-d372948383ff","Type":"ContainerStarted","Data":"7f1739c029eebd6fb961e5f23ff46342b147a2b89fb0eaec502375ac2e1fad8d"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.435551 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" event={"ID":"7bb6de6b-100d-4acf-8d51-d372948383ff","Type":"ContainerStarted","Data":"c756f2b354fc834e60df5aa78e9f46800fa2cb4c938413db96e794581d63f113"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.444173 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" event={"ID":"545a5916-91fb-4243-b825-fc09e352c3d5","Type":"ContainerStarted","Data":"ab00009daab5e9213d74c4ecc3b1b1ccc44589d9a462546a8a7179d89410f013"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.444227 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-sxpvs" event={"ID":"545a5916-91fb-4243-b825-fc09e352c3d5","Type":"ContainerStarted","Data":"0005adce14c967f21bb4a18dc2fafc5b5e44ac4d093f11f8d0e488674255324b"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.450532 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rvspl" event={"ID":"28beec1e-6abb-40ef-8d38-d567a20f9dd8","Type":"ContainerStarted","Data":"f7e944f6e006eaf619cec2895e4ecd148958fa5f6c3e86defe462bc0013d1112"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.456644 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-stkz9" event={"ID":"e308a4a1-6626-42c8-9e3b-8823beb2f204","Type":"ContainerStarted","Data":"e5ca5b0f2291b85ba71c91e7fd9e73331c0d74ee8966db54c8ababc3818b39d0"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.456702 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-stkz9" event={"ID":"e308a4a1-6626-42c8-9e3b-8823beb2f204","Type":"ContainerStarted","Data":"f60429a67be889900abc9adb00dc1c943598a2a3b5e34015973ff97ed860b51c"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.456830 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-stkz9" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.458073 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" podStartSLOduration=127.458061759 podStartE2EDuration="2m7.458061759s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.456407862 +0000 UTC m=+148.140364289" watchObservedRunningTime="2025-11-25 09:50:51.458061759 +0000 UTC m=+148.142018176" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.462857 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.463070 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.96303961 +0000 UTC m=+148.646996197 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.463229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.463843 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:51.963829443 +0000 UTC m=+148.647785870 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.465457 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-srlnm" event={"ID":"e3eb25a8-17ed-45b5-a875-42b9d33c54e7","Type":"ContainerStarted","Data":"319a1ca6ca25a37a1b31b7201582c0c898f78a8fafc66ddf3144356d28cef95f"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.475057 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" event={"ID":"b3716ff0-7157-4406-be92-89893b24a9ef","Type":"ContainerStarted","Data":"16969bd58716a1c65698d27d5fcdfc53f88d449b69ea4b63f1c985083496daa5"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.475131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" event={"ID":"b3716ff0-7157-4406-be92-89893b24a9ef","Type":"ContainerStarted","Data":"d7b3b05bba9fccd07ea3c73f7d8a4668b6cba6ec61cb2123c70d76ba1a755f07"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.490208 4926 generic.go:334] "Generic (PLEG): container finished" podID="e83f0f60-c4e8-43b1-9949-4aef1eb65994" containerID="3d578a1522e352b45bcf0e4e07b6a5d713dab080be1288c60197747cb6a3fd65" exitCode=0 Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.490333 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bd494" event={"ID":"e83f0f60-c4e8-43b1-9949-4aef1eb65994","Type":"ContainerStarted","Data":"f97f6f26c8f5a95bb0ee87aadd36679ee0e3d3272bbe0b3f64bd2ffe1d90a8eb"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.490437 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bd494" event={"ID":"e83f0f60-c4e8-43b1-9949-4aef1eb65994","Type":"ContainerStarted","Data":"786dd75cd375ec5484cda477a0c5244a96a4a1db980d0dc440afbfe0b049d388"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.490452 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bd494" event={"ID":"e83f0f60-c4e8-43b1-9949-4aef1eb65994","Type":"ContainerDied","Data":"3d578a1522e352b45bcf0e4e07b6a5d713dab080be1288c60197747cb6a3fd65"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.501967 4926 generic.go:334] "Generic (PLEG): container finished" podID="0c75d505-4bd3-45be-a65a-e532e517449d" containerID="c7fd6c02cc14890c03a5dbd71740ef616e3294e98b189d2038df1ec6515d2dcb" exitCode=0 Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.502078 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" event={"ID":"0c75d505-4bd3-45be-a65a-e532e517449d","Type":"ContainerStarted","Data":"7d6379df4be0fb2ec6b0d18159ee7b7a3fb0ccc24f048b02dc44804be45d9113"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.502144 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" event={"ID":"0c75d505-4bd3-45be-a65a-e532e517449d","Type":"ContainerDied","Data":"c7fd6c02cc14890c03a5dbd71740ef616e3294e98b189d2038df1ec6515d2dcb"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.502178 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.509785 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-v7qs6" event={"ID":"b2a3d4de-2095-4c4e-a490-6fc9f98dafa4","Type":"ContainerStarted","Data":"9ff54663b8bc4d969bc6cef72223fd9472a07e8e0fd78e47d6d2f3e86464848e"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.512734 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" event={"ID":"2ac8002f-f652-4341-93f7-ad4f828b564d","Type":"ContainerStarted","Data":"9811d3c46b761b7982ada30f1112b4922aa491bdca872edfc693841c202f222e"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.520537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-jq954" event={"ID":"577e54e1-5224-444f-823f-91d4d9f816c1","Type":"ContainerStarted","Data":"d392e30d565d56b1577a317ccc3ca94f1f10d1096199193dbbf34a881aae5892"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.532978 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rvspl" podStartSLOduration=8.532950572 podStartE2EDuration="8.532950572s" podCreationTimestamp="2025-11-25 09:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.532679014 +0000 UTC m=+148.216635431" watchObservedRunningTime="2025-11-25 09:50:51.532950572 +0000 UTC m=+148.216906999" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.534754 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-6xzbk" podStartSLOduration=127.534742693 podStartE2EDuration="2m7.534742693s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.506359078 +0000 UTC m=+148.190315505" watchObservedRunningTime="2025-11-25 09:50:51.534742693 +0000 UTC m=+148.218699120" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.535907 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" event={"ID":"7bc1de70-8871-4225-8a3e-d6f849fc864a","Type":"ContainerStarted","Data":"e8aa38d0c206c2096a42aa21dd8e0c4518ae96f2f07589dda3be3e0eb0f7fa86"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.537003 4926 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-ht99z container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" start-of-body= Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.537044 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.27:6443/healthz\": dial tcp 10.217.0.27:6443: connect: connection refused" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.541793 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" event={"ID":"b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c","Type":"ContainerStarted","Data":"832eee9626c25471baf6ad8faae464532cf475f75c98b94490e8124376d05c35"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.546481 4926 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-cbhpl container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" start-of-body= Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.546546 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" podUID="b7fcee12-0e8a-490e-ab21-0c6b8c90cc1c" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.36:8443/healthz\": dial tcp 10.217.0.36:8443: connect: connection refused" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.557118 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" event={"ID":"dd0f7444-df82-40cd-aaa3-60250afa1f36","Type":"ContainerStarted","Data":"8c6543c4e60c937ae9f30ace8193728c6107dd73a19f107096f7dd06a4d274ef"} Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.557640 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.557692 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.557957 4926 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-nn4ms container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.557975 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.22:8080/healthz\": dial tcp 10.217.0.22:8080: connect: connection refused" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.564436 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.564907 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.064887697 +0000 UTC m=+148.748844124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.570852 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-t9spw" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.624915 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" podStartSLOduration=127.624890388 podStartE2EDuration="2m7.624890388s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.580540401 +0000 UTC m=+148.264496828" watchObservedRunningTime="2025-11-25 09:50:51.624890388 +0000 UTC m=+148.308846815" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.626479 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-stkz9" podStartSLOduration=8.626471663 podStartE2EDuration="8.626471663s" podCreationTimestamp="2025-11-25 09:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.623804907 +0000 UTC m=+148.307761334" watchObservedRunningTime="2025-11-25 09:50:51.626471663 +0000 UTC m=+148.310428100" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.659482 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-68sgp" podStartSLOduration=127.659462738 podStartE2EDuration="2m7.659462738s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.658273684 +0000 UTC m=+148.342230111" watchObservedRunningTime="2025-11-25 09:50:51.659462738 +0000 UTC m=+148.343419165" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.667590 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.684810 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.184794836 +0000 UTC m=+148.868751263 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.765342 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-bd494" podStartSLOduration=127.765296287 podStartE2EDuration="2m7.765296287s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:51.720888799 +0000 UTC m=+148.404845226" watchObservedRunningTime="2025-11-25 09:50:51.765296287 +0000 UTC m=+148.449252714" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.769475 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.769760 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.269729623 +0000 UTC m=+148.953686050 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.769969 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.770511 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.270497255 +0000 UTC m=+148.954453682 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.870710 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.871186 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.371168698 +0000 UTC m=+149.055125125 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.958986 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5nkct"] Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.976369 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.983754 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.983818 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.983920 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.983978 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.985699 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:51 crc kubenswrapper[4926]: E1125 09:50:51.985997 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.485963612 +0000 UTC m=+149.169920039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.991739 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:51 crc kubenswrapper[4926]: I1125 09:50:51.994822 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.000247 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.029848 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5nkct"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.085507 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.085831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-utilities\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.085902 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-catalog-content\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.085953 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.086099 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h99jg\" (UniqueName: \"kubernetes.io/projected/85c11326-80b9-4be7-beda-d9ef3be2cd8a-kube-api-access-h99jg\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.086176 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.586137811 +0000 UTC m=+149.270094248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.094011 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.112315 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tlltp"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.113813 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.116982 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.140611 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tlltp"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.190925 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h99jg\" (UniqueName: \"kubernetes.io/projected/85c11326-80b9-4be7-beda-d9ef3be2cd8a-kube-api-access-h99jg\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.191002 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.191033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-utilities\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.191075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-catalog-content\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.191621 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-catalog-content\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.192111 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.692090245 +0000 UTC m=+149.376046672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.192581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-utilities\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.226620 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h99jg\" (UniqueName: \"kubernetes.io/projected/85c11326-80b9-4be7-beda-d9ef3be2cd8a-kube-api-access-h99jg\") pod \"certified-operators-5nkct\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.247471 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.250341 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:52 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:52 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:52 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.250412 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.261429 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.282232 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.293106 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.293329 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.793289223 +0000 UTC m=+149.477245650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.293426 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-catalog-content\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.293634 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.293708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-utilities\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.293735 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r99kv\" (UniqueName: \"kubernetes.io/projected/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-kube-api-access-r99kv\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.294644 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.794632771 +0000 UTC m=+149.478589198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.296993 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-brdt4"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.298257 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.319983 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brdt4"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.323579 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.395723 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.396107 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-utilities\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.396170 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r99kv\" (UniqueName: \"kubernetes.io/projected/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-kube-api-access-r99kv\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.396284 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-catalog-content\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.396883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-utilities\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.397335 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.897290691 +0000 UTC m=+149.581247108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.397445 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.397862 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:52.897853247 +0000 UTC m=+149.581809674 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.398070 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-catalog-content\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.427360 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r99kv\" (UniqueName: \"kubernetes.io/projected/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-kube-api-access-r99kv\") pod \"community-operators-tlltp\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.444259 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.499416 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.499838 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-utilities\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.499921 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-catalog-content\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.499968 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhk6k\" (UniqueName: \"kubernetes.io/projected/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-kube-api-access-jhk6k\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.500151 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.000128546 +0000 UTC m=+149.684084983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.502459 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n4xnt"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.503768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.553038 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n4xnt"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.600695 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" event={"ID":"2ac8002f-f652-4341-93f7-ad4f828b564d","Type":"ContainerStarted","Data":"549987dff87ec4e72c73e17f83c9e118189662b75353872bf344f68fb8dcb077"} Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603281 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-utilities\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603337 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603368 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-utilities\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603396 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cntdj\" (UniqueName: \"kubernetes.io/projected/f78474ad-ae46-4070-9049-bfd94d411642-kube-api-access-cntdj\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603442 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-catalog-content\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603461 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhk6k\" (UniqueName: \"kubernetes.io/projected/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-kube-api-access-jhk6k\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.603487 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-catalog-content\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.603791 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.103777664 +0000 UTC m=+149.787734091 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.609373 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-catalog-content\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.611375 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-utilities\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.611674 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.611714 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.626955 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.635798 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhk6k\" (UniqueName: \"kubernetes.io/projected/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-kube-api-access-jhk6k\") pod \"certified-operators-brdt4\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.649354 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.650127 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.668192 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-cbhpl" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.668461 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.668793 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.669456 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.677276 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.711635 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.715490 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.215443319 +0000 UTC m=+149.899399776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.716137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-catalog-content\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.718212 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-catalog-content\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.732868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-utilities\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.732975 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.733347 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cntdj\" (UniqueName: \"kubernetes.io/projected/f78474ad-ae46-4070-9049-bfd94d411642-kube-api-access-cntdj\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.783610 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.28358581 +0000 UTC m=+149.967542237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.795087 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-utilities\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.800817 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cntdj\" (UniqueName: \"kubernetes.io/projected/f78474ad-ae46-4070-9049-bfd94d411642-kube-api-access-cntdj\") pod \"community-operators-n4xnt\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.835544 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.835951 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.835982 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.836495 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.336476199 +0000 UTC m=+150.020432626 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.855026 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.899467 4926 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.937980 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.938028 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.938044 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:52 crc kubenswrapper[4926]: E1125 09:50:52.938400 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.438384148 +0000 UTC m=+150.122340575 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.938462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:52 crc kubenswrapper[4926]: I1125 09:50:52.961559 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.025450 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.038523 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:53 crc kubenswrapper[4926]: E1125 09:50:53.038869 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.538847285 +0000 UTC m=+150.222803712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.140023 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:53 crc kubenswrapper[4926]: E1125 09:50:53.140514 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-25 09:50:53.640495937 +0000 UTC m=+150.324452364 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-8zltj" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 25 09:50:53 crc kubenswrapper[4926]: W1125 09:50:53.184410 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-caa0fe4cdd1ba01166abc88844f15b11bb640d1053f67afb0a853c9466e0a5d8 WatchSource:0}: Error finding container caa0fe4cdd1ba01166abc88844f15b11bb640d1053f67afb0a853c9466e0a5d8: Status 404 returned error can't find the container with id caa0fe4cdd1ba01166abc88844f15b11bb640d1053f67afb0a853c9466e0a5d8 Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.210360 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-brdt4"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.214663 4926 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-25T09:50:52.899506646Z","Handler":null,"Name":""} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.235383 4926 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.235412 4926 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.241691 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.274074 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5nkct"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.279073 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:53 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:53 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:53 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.279136 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.280020 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.329109 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tlltp"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.346279 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.414275 4926 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.414328 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.463856 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n4xnt"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.469785 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.544969 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-8zltj\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.591594 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.621447 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerStarted","Data":"bbdba7e97b417d3b64f9a02d0ff407de603607cffcc82fcb731f68b75900a07f"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.624671 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224","Type":"ContainerStarted","Data":"cfb4d2494cd2e10f87ad6658b1e7b49976ea12fd3b87cc4e65a767272f7ced81"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.633892 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"70bb255f8fac2bb81646f591c7cc7cc3637b2f25c6887de8f21bc88eb7b4eb30"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.633973 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"caa0fe4cdd1ba01166abc88844f15b11bb640d1053f67afb0a853c9466e0a5d8"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.647272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerStarted","Data":"c8ae29ea7608ef84a658a9ea87e7da600d1b899ae67a9e0ca2e888391a2d4b1f"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.649173 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerStarted","Data":"ba5bc90ff67b67f201fe37455c9fe32c8e20e882768b124c708f505c300fa329"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.650830 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5e7f7729c35c0fae827e2b23c4cd82b5f3510e4cd905727132118623490b7d8e"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.663449 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" event={"ID":"2ac8002f-f652-4341-93f7-ad4f828b564d","Type":"ContainerStarted","Data":"ba270e44906171a09d4b16d202e4629711cd380168fb771b172f9adda763cc95"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.667061 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5nkct" event={"ID":"85c11326-80b9-4be7-beda-d9ef3be2cd8a","Type":"ContainerStarted","Data":"b543467a386d2390b41600a897ec19e6dcf141f2e0d8d806f5e1fec6221d2ce1"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.683623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"ab13e54790e4f4224c3ff8f96d8f258d448cb51b2f0d8cf633ced21e90e6f4bd"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.683670 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a925a1814324e852fca2cab28552d71f3af3cab9a1b206d8c99a80694f2583c6"} Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.889122 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2fxvt"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.890590 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.965494 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2fxvt"] Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.979290 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.980222 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-utilities\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.980361 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-catalog-content\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:53 crc kubenswrapper[4926]: I1125 09:50:53.980433 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c84n7\" (UniqueName: \"kubernetes.io/projected/196347d9-90c7-41e5-a8c5-75acad8f047a-kube-api-access-c84n7\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.010655 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zltj"] Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.023127 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.081196 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-utilities\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.081264 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-catalog-content\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.081316 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c84n7\" (UniqueName: \"kubernetes.io/projected/196347d9-90c7-41e5-a8c5-75acad8f047a-kube-api-access-c84n7\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.082117 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-utilities\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.082289 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-catalog-content\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.105741 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c84n7\" (UniqueName: \"kubernetes.io/projected/196347d9-90c7-41e5-a8c5-75acad8f047a-kube-api-access-c84n7\") pod \"redhat-marketplace-2fxvt\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.114228 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.249222 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:54 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:54 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:54 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.249295 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.289451 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b8bvj"] Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.297394 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8bvj"] Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.297550 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.415521 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zprk\" (UniqueName: \"kubernetes.io/projected/39af42eb-1c94-4a21-87bb-a370e45a3f1d-kube-api-access-4zprk\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.415593 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-catalog-content\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.415587 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2fxvt"] Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.415704 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-utilities\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.516963 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-utilities\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.517059 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zprk\" (UniqueName: \"kubernetes.io/projected/39af42eb-1c94-4a21-87bb-a370e45a3f1d-kube-api-access-4zprk\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.517086 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-catalog-content\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.517580 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-catalog-content\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.518272 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-utilities\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.538501 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zprk\" (UniqueName: \"kubernetes.io/projected/39af42eb-1c94-4a21-87bb-a370e45a3f1d-kube-api-access-4zprk\") pod \"redhat-marketplace-b8bvj\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.633014 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.690521 4926 generic.go:334] "Generic (PLEG): container finished" podID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerID="0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8" exitCode=0 Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.690578 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerDied","Data":"0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.690897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerStarted","Data":"51fe78f7135327e3d4884b7ad3ebd62ba77833cb1bc5f16eaee0b1ad7a32486c"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.694246 4926 generic.go:334] "Generic (PLEG): container finished" podID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerID="edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d" exitCode=0 Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.694313 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerDied","Data":"edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.695183 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.697716 4926 generic.go:334] "Generic (PLEG): container finished" podID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerID="c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860" exitCode=0 Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.697773 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerDied","Data":"c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.700267 4926 generic.go:334] "Generic (PLEG): container finished" podID="dd0f7444-df82-40cd-aaa3-60250afa1f36" containerID="8c6543c4e60c937ae9f30ace8193728c6107dd73a19f107096f7dd06a4d274ef" exitCode=0 Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.700312 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" event={"ID":"dd0f7444-df82-40cd-aaa3-60250afa1f36","Type":"ContainerDied","Data":"8c6543c4e60c937ae9f30ace8193728c6107dd73a19f107096f7dd06a4d274ef"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.706925 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224","Type":"ContainerStarted","Data":"beb88795ac2232abf58c1788a2cc87c28f7546a98eb1995a476a00293bd71b97"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.709442 4926 generic.go:334] "Generic (PLEG): container finished" podID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerID="157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d" exitCode=0 Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.709509 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5nkct" event={"ID":"85c11326-80b9-4be7-beda-d9ef3be2cd8a","Type":"ContainerDied","Data":"157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.723179 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6956cb7a3b8feb64c4b7e3c4a2c6802c8cbb9b3c72d923cc726fe838aee1b666"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.723327 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.725164 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78474ad-ae46-4070-9049-bfd94d411642" containerID="7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4" exitCode=0 Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.725229 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerDied","Data":"7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.774442 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" event={"ID":"2ac8002f-f652-4341-93f7-ad4f828b564d","Type":"ContainerStarted","Data":"4e88dda12ef1d367cba7ff3ca96db8b46d7cbd4b7c51cb84bec64ab27242e937"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.776857 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" event={"ID":"36216347-d1f5-4db3-8bb8-6654caba4839","Type":"ContainerStarted","Data":"7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.776906 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" event={"ID":"36216347-d1f5-4db3-8bb8-6654caba4839","Type":"ContainerStarted","Data":"31b51098160244cad5ffd4c299c03bc7dd70de57e15783e53acade310ff42fe8"} Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.777096 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.843820 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.843790935 podStartE2EDuration="2.843790935s" podCreationTimestamp="2025-11-25 09:50:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:54.834274255 +0000 UTC m=+151.518230682" watchObservedRunningTime="2025-11-25 09:50:54.843790935 +0000 UTC m=+151.527747362" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.892053 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" podStartSLOduration=130.892029882 podStartE2EDuration="2m10.892029882s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:54.883966723 +0000 UTC m=+151.567923160" watchObservedRunningTime="2025-11-25 09:50:54.892029882 +0000 UTC m=+151.575986309" Nov 25 09:50:54 crc kubenswrapper[4926]: I1125 09:50:54.909308 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-hw4f2" podStartSLOduration=11.909285791 podStartE2EDuration="11.909285791s" podCreationTimestamp="2025-11-25 09:50:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:50:54.905304088 +0000 UTC m=+151.589260515" watchObservedRunningTime="2025-11-25 09:50:54.909285791 +0000 UTC m=+151.593242238" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.087585 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8bvj"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.094145 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qs6rc"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.095663 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.100427 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qs6rc"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.103080 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.129493 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.130804 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.134136 4926 patch_prober.go:28] interesting pod/console-f9d7485db-hfbwq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.134204 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-hfbwq" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.232169 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgtcx\" (UniqueName: \"kubernetes.io/projected/d82538b8-f059-4fc6-bd88-622804100131-kube-api-access-hgtcx\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.232416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-utilities\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.233172 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-catalog-content\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.248672 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:55 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:55 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:55 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.249167 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.290453 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-kjdgl" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.333789 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgtcx\" (UniqueName: \"kubernetes.io/projected/d82538b8-f059-4fc6-bd88-622804100131-kube-api-access-hgtcx\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.334202 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-utilities\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.334348 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-catalog-content\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.335023 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-catalog-content\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.335155 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-utilities\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.362998 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgtcx\" (UniqueName: \"kubernetes.io/projected/d82538b8-f059-4fc6-bd88-622804100131-kube-api-access-hgtcx\") pod \"redhat-operators-qs6rc\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.412579 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.505229 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jnhlw"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.507023 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.507801 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.508576 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.512689 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jnhlw"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.520235 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.520507 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.523010 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.539645 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gft47\" (UniqueName: \"kubernetes.io/projected/3b6717a6-c510-4800-bc0b-bf4506132cef-kube-api-access-gft47\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.539699 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-utilities\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.539756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.539954 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.540001 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-catalog-content\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.642488 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gft47\" (UniqueName: \"kubernetes.io/projected/3b6717a6-c510-4800-bc0b-bf4506132cef-kube-api-access-gft47\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.642532 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-utilities\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.642582 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.642627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.642656 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-catalog-content\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.643209 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-catalog-content\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.644149 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-utilities\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.644313 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.669724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gft47\" (UniqueName: \"kubernetes.io/projected/3b6717a6-c510-4800-bc0b-bf4506132cef-kube-api-access-gft47\") pod \"redhat-operators-jnhlw\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.670427 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.767742 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qs6rc"] Nov 25 09:50:55 crc kubenswrapper[4926]: W1125 09:50:55.778294 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd82538b8_f059_4fc6_bd88_622804100131.slice/crio-28e5f3f020a1e3a411a49bde3fde59600c26a4202688235ff35d7729d277fcf8 WatchSource:0}: Error finding container 28e5f3f020a1e3a411a49bde3fde59600c26a4202688235ff35d7729d277fcf8: Status 404 returned error can't find the container with id 28e5f3f020a1e3a411a49bde3fde59600c26a4202688235ff35d7729d277fcf8 Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.788791 4926 generic.go:334] "Generic (PLEG): container finished" podID="9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224" containerID="beb88795ac2232abf58c1788a2cc87c28f7546a98eb1995a476a00293bd71b97" exitCode=0 Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.789359 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224","Type":"ContainerDied","Data":"beb88795ac2232abf58c1788a2cc87c28f7546a98eb1995a476a00293bd71b97"} Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.796796 4926 generic.go:334] "Generic (PLEG): container finished" podID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerID="d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28" exitCode=0 Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.797616 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8bvj" event={"ID":"39af42eb-1c94-4a21-87bb-a370e45a3f1d","Type":"ContainerDied","Data":"d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28"} Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.797706 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8bvj" event={"ID":"39af42eb-1c94-4a21-87bb-a370e45a3f1d","Type":"ContainerStarted","Data":"7d0fc311be6746642bf6f05a207ba2c9a8679c2517d8705dc583b71e16deef8c"} Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.835013 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:50:55 crc kubenswrapper[4926]: I1125 09:50:55.854730 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.039477 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.039540 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.047694 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.098512 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.162021 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.162082 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.162793 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.162873 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.185682 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jnhlw"] Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.203602 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.246693 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.249220 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:56 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:56 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:56 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.249279 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.258396 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd0f7444-df82-40cd-aaa3-60250afa1f36-config-volume\") pod \"dd0f7444-df82-40cd-aaa3-60250afa1f36\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.258735 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd0f7444-df82-40cd-aaa3-60250afa1f36-secret-volume\") pod \"dd0f7444-df82-40cd-aaa3-60250afa1f36\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.258793 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmj57\" (UniqueName: \"kubernetes.io/projected/dd0f7444-df82-40cd-aaa3-60250afa1f36-kube-api-access-gmj57\") pod \"dd0f7444-df82-40cd-aaa3-60250afa1f36\" (UID: \"dd0f7444-df82-40cd-aaa3-60250afa1f36\") " Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.259393 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd0f7444-df82-40cd-aaa3-60250afa1f36-config-volume" (OuterVolumeSpecName: "config-volume") pod "dd0f7444-df82-40cd-aaa3-60250afa1f36" (UID: "dd0f7444-df82-40cd-aaa3-60250afa1f36"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.267744 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd0f7444-df82-40cd-aaa3-60250afa1f36-kube-api-access-gmj57" (OuterVolumeSpecName: "kube-api-access-gmj57") pod "dd0f7444-df82-40cd-aaa3-60250afa1f36" (UID: "dd0f7444-df82-40cd-aaa3-60250afa1f36"). InnerVolumeSpecName "kube-api-access-gmj57". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.282309 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd0f7444-df82-40cd-aaa3-60250afa1f36-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "dd0f7444-df82-40cd-aaa3-60250afa1f36" (UID: "dd0f7444-df82-40cd-aaa3-60250afa1f36"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.304093 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.362566 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/dd0f7444-df82-40cd-aaa3-60250afa1f36-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.362705 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmj57\" (UniqueName: \"kubernetes.io/projected/dd0f7444-df82-40cd-aaa3-60250afa1f36-kube-api-access-gmj57\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.363404 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/dd0f7444-df82-40cd-aaa3-60250afa1f36-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.449415 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.811723 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca","Type":"ContainerStarted","Data":"e5104ffebeedc43b48457337c5624e7e96fc9e5260f47ce671b04c0988a46eea"} Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.820459 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.820913 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq" event={"ID":"dd0f7444-df82-40cd-aaa3-60250afa1f36","Type":"ContainerDied","Data":"091b294a5025334a1f56b13080a5d8c9553d0b993299176d05642e1302748036"} Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.821049 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="091b294a5025334a1f56b13080a5d8c9553d0b993299176d05642e1302748036" Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.829335 4926 generic.go:334] "Generic (PLEG): container finished" podID="d82538b8-f059-4fc6-bd88-622804100131" containerID="be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa" exitCode=0 Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.829398 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs6rc" event={"ID":"d82538b8-f059-4fc6-bd88-622804100131","Type":"ContainerDied","Data":"be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa"} Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.829429 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs6rc" event={"ID":"d82538b8-f059-4fc6-bd88-622804100131","Type":"ContainerStarted","Data":"28e5f3f020a1e3a411a49bde3fde59600c26a4202688235ff35d7729d277fcf8"} Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.838697 4926 generic.go:334] "Generic (PLEG): container finished" podID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerID="16c3546f860b1a69781a75d691199e14a76dbc859c049baa934df377b91763c0" exitCode=0 Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.840660 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnhlw" event={"ID":"3b6717a6-c510-4800-bc0b-bf4506132cef","Type":"ContainerDied","Data":"16c3546f860b1a69781a75d691199e14a76dbc859c049baa934df377b91763c0"} Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.840704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnhlw" event={"ID":"3b6717a6-c510-4800-bc0b-bf4506132cef","Type":"ContainerStarted","Data":"51ad92ddd9840ab5e1fde4f49f4fc795239dca09022f5064a041da1df87f06af"} Nov 25 09:50:56 crc kubenswrapper[4926]: I1125 09:50:56.847552 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-bd494" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.253011 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:57 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:57 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:57 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.253336 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.436352 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.585682 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kube-api-access\") pod \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.585825 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kubelet-dir\") pod \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\" (UID: \"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224\") " Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.586141 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224" (UID: "9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.621412 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224" (UID: "9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.687397 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.687432 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.853045 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224","Type":"ContainerDied","Data":"cfb4d2494cd2e10f87ad6658b1e7b49976ea12fd3b87cc4e65a767272f7ced81"} Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.853368 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfb4d2494cd2e10f87ad6658b1e7b49976ea12fd3b87cc4e65a767272f7ced81" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.853466 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.861994 4926 generic.go:334] "Generic (PLEG): container finished" podID="0414cbf3-5dca-4fb0-8e0c-093b7f9026ca" containerID="542535bcef7940dfd3fe7b4029ec78bd483fc12af3f956bcd5541495e3a4eeb5" exitCode=0 Nov 25 09:50:57 crc kubenswrapper[4926]: I1125 09:50:57.863705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca","Type":"ContainerDied","Data":"542535bcef7940dfd3fe7b4029ec78bd483fc12af3f956bcd5541495e3a4eeb5"} Nov 25 09:50:58 crc kubenswrapper[4926]: I1125 09:50:58.249344 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:58 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:58 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:58 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:58 crc kubenswrapper[4926]: I1125 09:50:58.249410 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.249579 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:50:59 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:50:59 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:50:59 crc kubenswrapper[4926]: healthz check failed Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.249657 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.255256 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.425304 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kubelet-dir\") pod \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.425560 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kube-api-access\") pod \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\" (UID: \"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca\") " Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.426608 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0414cbf3-5dca-4fb0-8e0c-093b7f9026ca" (UID: "0414cbf3-5dca-4fb0-8e0c-093b7f9026ca"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.431193 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0414cbf3-5dca-4fb0-8e0c-093b7f9026ca" (UID: "0414cbf3-5dca-4fb0-8e0c-093b7f9026ca"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.527532 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.527575 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0414cbf3-5dca-4fb0-8e0c-093b7f9026ca-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.890801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"0414cbf3-5dca-4fb0-8e0c-093b7f9026ca","Type":"ContainerDied","Data":"e5104ffebeedc43b48457337c5624e7e96fc9e5260f47ce671b04c0988a46eea"} Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.891100 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5104ffebeedc43b48457337c5624e7e96fc9e5260f47ce671b04c0988a46eea" Nov 25 09:50:59 crc kubenswrapper[4926]: I1125 09:50:59.890879 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 25 09:51:00 crc kubenswrapper[4926]: I1125 09:51:00.249395 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:00 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:00 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:00 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:00 crc kubenswrapper[4926]: I1125 09:51:00.249459 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:01 crc kubenswrapper[4926]: I1125 09:51:01.104542 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-stkz9" Nov 25 09:51:01 crc kubenswrapper[4926]: I1125 09:51:01.251646 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:01 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:01 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:01 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:01 crc kubenswrapper[4926]: I1125 09:51:01.251729 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:02 crc kubenswrapper[4926]: I1125 09:51:02.250005 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:02 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:02 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:02 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:02 crc kubenswrapper[4926]: I1125 09:51:02.250360 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:03 crc kubenswrapper[4926]: I1125 09:51:03.248916 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:03 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:03 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:03 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:03 crc kubenswrapper[4926]: I1125 09:51:03.249019 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:04 crc kubenswrapper[4926]: I1125 09:51:04.248508 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:04 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:04 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:04 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:04 crc kubenswrapper[4926]: I1125 09:51:04.248903 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:05 crc kubenswrapper[4926]: I1125 09:51:05.119968 4926 patch_prober.go:28] interesting pod/console-f9d7485db-hfbwq container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" start-of-body= Nov 25 09:51:05 crc kubenswrapper[4926]: I1125 09:51:05.120046 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-hfbwq" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.30:8443/health\": dial tcp 10.217.0.30:8443: connect: connection refused" Nov 25 09:51:05 crc kubenswrapper[4926]: I1125 09:51:05.247138 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:05 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:05 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:05 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:05 crc kubenswrapper[4926]: I1125 09:51:05.247218 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.162495 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.162857 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.162512 4926 patch_prober.go:28] interesting pod/downloads-7954f5f757-dnjv7 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.162976 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dnjv7" podUID="53637ef2-e0c4-4e94-aca6-44b2d4bfc500" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.32:8080/\": dial tcp 10.217.0.32:8080: connect: connection refused" Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.248042 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:06 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:06 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:06 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.248125 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.455243 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.464482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/81f30c43-dacf-407f-adb3-bdbae866485f-metrics-certs\") pod \"network-metrics-daemon-xpn28\" (UID: \"81f30c43-dacf-407f-adb3-bdbae866485f\") " pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:51:06 crc kubenswrapper[4926]: I1125 09:51:06.633861 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-xpn28" Nov 25 09:51:07 crc kubenswrapper[4926]: I1125 09:51:07.248995 4926 patch_prober.go:28] interesting pod/router-default-5444994796-bngtn container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 25 09:51:07 crc kubenswrapper[4926]: [-]has-synced failed: reason withheld Nov 25 09:51:07 crc kubenswrapper[4926]: [+]process-running ok Nov 25 09:51:07 crc kubenswrapper[4926]: healthz check failed Nov 25 09:51:07 crc kubenswrapper[4926]: I1125 09:51:07.249070 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-bngtn" podUID="7a29e45e-3063-4a6d-b215-0765aa2df10a" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 25 09:51:08 crc kubenswrapper[4926]: I1125 09:51:08.249174 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:51:08 crc kubenswrapper[4926]: I1125 09:51:08.251931 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-bngtn" Nov 25 09:51:13 crc kubenswrapper[4926]: I1125 09:51:13.598447 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:51:14 crc kubenswrapper[4926]: I1125 09:51:14.015174 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:51:14 crc kubenswrapper[4926]: I1125 09:51:14.015281 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:51:15 crc kubenswrapper[4926]: I1125 09:51:15.216511 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:51:15 crc kubenswrapper[4926]: I1125 09:51:15.222585 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 09:51:16 crc kubenswrapper[4926]: I1125 09:51:16.170948 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dnjv7" Nov 25 09:51:26 crc kubenswrapper[4926]: I1125 09:51:26.318432 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jsgtd" Nov 25 09:51:28 crc kubenswrapper[4926]: E1125 09:51:28.526619 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 09:51:28 crc kubenswrapper[4926]: E1125 09:51:28.526831 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4zprk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-b8bvj_openshift-marketplace(39af42eb-1c94-4a21-87bb-a370e45a3f1d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:28 crc kubenswrapper[4926]: E1125 09:51:28.528004 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-b8bvj" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.289614 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-b8bvj" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.452877 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.453337 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c84n7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-2fxvt_openshift-marketplace(196347d9-90c7-41e5-a8c5-75acad8f047a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.454751 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-2fxvt" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.467139 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.467402 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gft47,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jnhlw_openshift-marketplace(3b6717a6-c510-4800-bc0b-bf4506132cef): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:31 crc kubenswrapper[4926]: E1125 09:51:31.468904 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jnhlw" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" Nov 25 09:51:32 crc kubenswrapper[4926]: I1125 09:51:32.253268 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 25 09:51:32 crc kubenswrapper[4926]: E1125 09:51:32.667638 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jnhlw" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" Nov 25 09:51:32 crc kubenswrapper[4926]: E1125 09:51:32.667666 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-2fxvt" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" Nov 25 09:51:32 crc kubenswrapper[4926]: E1125 09:51:32.744564 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 09:51:32 crc kubenswrapper[4926]: E1125 09:51:32.744798 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h99jg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-5nkct_openshift-marketplace(85c11326-80b9-4be7-beda-d9ef3be2cd8a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:32 crc kubenswrapper[4926]: E1125 09:51:32.746002 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-5nkct" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.027161 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-5nkct" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.122774 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.123079 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cntdj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-n4xnt_openshift-marketplace(f78474ad-ae46-4070-9049-bfd94d411642): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.126030 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-n4xnt" podUID="f78474ad-ae46-4070-9049-bfd94d411642" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.168109 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.168299 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r99kv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-tlltp_openshift-marketplace(7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.169488 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-tlltp" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.196688 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.196963 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jhk6k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-brdt4_openshift-marketplace(78f5ad10-72d9-4dff-b9c4-54b9bfe33958): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.199715 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-brdt4" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.297086 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.298057 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hgtcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-qs6rc_openshift-marketplace(d82538b8-f059-4fc6-bd88-622804100131): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 25 09:51:34 crc kubenswrapper[4926]: E1125 09:51:34.299336 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-qs6rc" podUID="d82538b8-f059-4fc6-bd88-622804100131" Nov 25 09:51:34 crc kubenswrapper[4926]: I1125 09:51:34.418525 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-xpn28"] Nov 25 09:51:34 crc kubenswrapper[4926]: W1125 09:51:34.429326 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81f30c43_dacf_407f_adb3_bdbae866485f.slice/crio-276d9f775fe8dc408e6c684a01d36f5f6609454f8d1bfa9553817be913576848 WatchSource:0}: Error finding container 276d9f775fe8dc408e6c684a01d36f5f6609454f8d1bfa9553817be913576848: Status 404 returned error can't find the container with id 276d9f775fe8dc408e6c684a01d36f5f6609454f8d1bfa9553817be913576848 Nov 25 09:51:35 crc kubenswrapper[4926]: I1125 09:51:35.148118 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xpn28" event={"ID":"81f30c43-dacf-407f-adb3-bdbae866485f","Type":"ContainerStarted","Data":"5119a0bacacaf1ec7cf315b26841464a9b5d354f36313ae28ba256a7db3ba214"} Nov 25 09:51:35 crc kubenswrapper[4926]: I1125 09:51:35.148546 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xpn28" event={"ID":"81f30c43-dacf-407f-adb3-bdbae866485f","Type":"ContainerStarted","Data":"276d9f775fe8dc408e6c684a01d36f5f6609454f8d1bfa9553817be913576848"} Nov 25 09:51:35 crc kubenswrapper[4926]: E1125 09:51:35.150202 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-tlltp" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" Nov 25 09:51:35 crc kubenswrapper[4926]: E1125 09:51:35.150539 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-brdt4" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" Nov 25 09:51:35 crc kubenswrapper[4926]: E1125 09:51:35.150585 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-qs6rc" podUID="d82538b8-f059-4fc6-bd88-622804100131" Nov 25 09:51:35 crc kubenswrapper[4926]: E1125 09:51:35.151153 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-n4xnt" podUID="f78474ad-ae46-4070-9049-bfd94d411642" Nov 25 09:51:36 crc kubenswrapper[4926]: I1125 09:51:36.155604 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-xpn28" event={"ID":"81f30c43-dacf-407f-adb3-bdbae866485f","Type":"ContainerStarted","Data":"260b1812adabcbba97c9acb41671c341b1c67691bf65a5f43bb3b011587061b4"} Nov 25 09:51:36 crc kubenswrapper[4926]: I1125 09:51:36.176949 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-xpn28" podStartSLOduration=172.176895425 podStartE2EDuration="2m52.176895425s" podCreationTimestamp="2025-11-25 09:48:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:51:36.171605115 +0000 UTC m=+192.855561552" watchObservedRunningTime="2025-11-25 09:51:36.176895425 +0000 UTC m=+192.860851852" Nov 25 09:51:37 crc kubenswrapper[4926]: I1125 09:51:37.876790 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht99z"] Nov 25 09:51:43 crc kubenswrapper[4926]: I1125 09:51:43.223300 4926 generic.go:334] "Generic (PLEG): container finished" podID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerID="51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd" exitCode=0 Nov 25 09:51:43 crc kubenswrapper[4926]: I1125 09:51:43.223383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8bvj" event={"ID":"39af42eb-1c94-4a21-87bb-a370e45a3f1d","Type":"ContainerDied","Data":"51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd"} Nov 25 09:51:44 crc kubenswrapper[4926]: I1125 09:51:44.014451 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:51:44 crc kubenswrapper[4926]: I1125 09:51:44.014866 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:51:45 crc kubenswrapper[4926]: I1125 09:51:45.238651 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8bvj" event={"ID":"39af42eb-1c94-4a21-87bb-a370e45a3f1d","Type":"ContainerStarted","Data":"1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037"} Nov 25 09:51:45 crc kubenswrapper[4926]: I1125 09:51:45.241324 4926 generic.go:334] "Generic (PLEG): container finished" podID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerID="da57a4320af15283c133274329225d2233e1191fe473be2cf6f92c81451e0410" exitCode=0 Nov 25 09:51:45 crc kubenswrapper[4926]: I1125 09:51:45.241386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnhlw" event={"ID":"3b6717a6-c510-4800-bc0b-bf4506132cef","Type":"ContainerDied","Data":"da57a4320af15283c133274329225d2233e1191fe473be2cf6f92c81451e0410"} Nov 25 09:51:45 crc kubenswrapper[4926]: I1125 09:51:45.262394 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b8bvj" podStartSLOduration=3.191226945 podStartE2EDuration="51.262371677s" podCreationTimestamp="2025-11-25 09:50:54 +0000 UTC" firstStartedPulling="2025-11-25 09:50:55.813923643 +0000 UTC m=+152.497880070" lastFinishedPulling="2025-11-25 09:51:43.885068385 +0000 UTC m=+200.569024802" observedRunningTime="2025-11-25 09:51:45.259134898 +0000 UTC m=+201.943091325" watchObservedRunningTime="2025-11-25 09:51:45.262371677 +0000 UTC m=+201.946328104" Nov 25 09:51:46 crc kubenswrapper[4926]: I1125 09:51:46.252548 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnhlw" event={"ID":"3b6717a6-c510-4800-bc0b-bf4506132cef","Type":"ContainerStarted","Data":"14169a6f0bb545959615df9c1036eaa2e37d5fc0e4647029bc26c5b82d2b1370"} Nov 25 09:51:46 crc kubenswrapper[4926]: I1125 09:51:46.274230 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jnhlw" podStartSLOduration=2.3989900410000002 podStartE2EDuration="51.274209842s" podCreationTimestamp="2025-11-25 09:50:55 +0000 UTC" firstStartedPulling="2025-11-25 09:50:56.842415845 +0000 UTC m=+153.526372272" lastFinishedPulling="2025-11-25 09:51:45.717635646 +0000 UTC m=+202.401592073" observedRunningTime="2025-11-25 09:51:46.271057095 +0000 UTC m=+202.955013532" watchObservedRunningTime="2025-11-25 09:51:46.274209842 +0000 UTC m=+202.958166259" Nov 25 09:51:47 crc kubenswrapper[4926]: I1125 09:51:47.272369 4926 generic.go:334] "Generic (PLEG): container finished" podID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerID="8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d" exitCode=0 Nov 25 09:51:47 crc kubenswrapper[4926]: I1125 09:51:47.272449 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5nkct" event={"ID":"85c11326-80b9-4be7-beda-d9ef3be2cd8a","Type":"ContainerDied","Data":"8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d"} Nov 25 09:51:49 crc kubenswrapper[4926]: I1125 09:51:49.284203 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5nkct" event={"ID":"85c11326-80b9-4be7-beda-d9ef3be2cd8a","Type":"ContainerStarted","Data":"50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba"} Nov 25 09:51:49 crc kubenswrapper[4926]: I1125 09:51:49.287606 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerStarted","Data":"c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10"} Nov 25 09:51:49 crc kubenswrapper[4926]: I1125 09:51:49.291221 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerStarted","Data":"6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346"} Nov 25 09:51:49 crc kubenswrapper[4926]: I1125 09:51:49.304506 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5nkct" podStartSLOduration=4.041313973 podStartE2EDuration="58.304489172s" podCreationTimestamp="2025-11-25 09:50:51 +0000 UTC" firstStartedPulling="2025-11-25 09:50:54.720221342 +0000 UTC m=+151.404177769" lastFinishedPulling="2025-11-25 09:51:48.983396541 +0000 UTC m=+205.667352968" observedRunningTime="2025-11-25 09:51:49.303413762 +0000 UTC m=+205.987370189" watchObservedRunningTime="2025-11-25 09:51:49.304489172 +0000 UTC m=+205.988445599" Nov 25 09:51:50 crc kubenswrapper[4926]: I1125 09:51:50.300310 4926 generic.go:334] "Generic (PLEG): container finished" podID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerID="c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10" exitCode=0 Nov 25 09:51:50 crc kubenswrapper[4926]: I1125 09:51:50.300376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerDied","Data":"c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10"} Nov 25 09:51:50 crc kubenswrapper[4926]: I1125 09:51:50.303976 4926 generic.go:334] "Generic (PLEG): container finished" podID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerID="6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346" exitCode=0 Nov 25 09:51:50 crc kubenswrapper[4926]: I1125 09:51:50.304049 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerDied","Data":"6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346"} Nov 25 09:51:50 crc kubenswrapper[4926]: I1125 09:51:50.317280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerStarted","Data":"41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2"} Nov 25 09:51:50 crc kubenswrapper[4926]: I1125 09:51:50.321862 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerStarted","Data":"b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8"} Nov 25 09:51:51 crc kubenswrapper[4926]: I1125 09:51:51.329751 4926 generic.go:334] "Generic (PLEG): container finished" podID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerID="41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2" exitCode=0 Nov 25 09:51:51 crc kubenswrapper[4926]: I1125 09:51:51.329852 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerDied","Data":"41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2"} Nov 25 09:51:51 crc kubenswrapper[4926]: I1125 09:51:51.334530 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78474ad-ae46-4070-9049-bfd94d411642" containerID="b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8" exitCode=0 Nov 25 09:51:51 crc kubenswrapper[4926]: I1125 09:51:51.334584 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerDied","Data":"b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8"} Nov 25 09:51:52 crc kubenswrapper[4926]: I1125 09:51:52.325828 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:51:52 crc kubenswrapper[4926]: I1125 09:51:52.325905 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:51:52 crc kubenswrapper[4926]: I1125 09:51:52.343158 4926 generic.go:334] "Generic (PLEG): container finished" podID="d82538b8-f059-4fc6-bd88-622804100131" containerID="116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514" exitCode=0 Nov 25 09:51:52 crc kubenswrapper[4926]: I1125 09:51:52.343223 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs6rc" event={"ID":"d82538b8-f059-4fc6-bd88-622804100131","Type":"ContainerDied","Data":"116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514"} Nov 25 09:51:52 crc kubenswrapper[4926]: I1125 09:51:52.846210 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:51:54 crc kubenswrapper[4926]: I1125 09:51:54.355450 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerStarted","Data":"63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691"} Nov 25 09:51:54 crc kubenswrapper[4926]: I1125 09:51:54.371523 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tlltp" podStartSLOduration=4.37614171 podStartE2EDuration="1m2.371442308s" podCreationTimestamp="2025-11-25 09:50:52 +0000 UTC" firstStartedPulling="2025-11-25 09:50:54.697542819 +0000 UTC m=+151.381499246" lastFinishedPulling="2025-11-25 09:51:52.692843417 +0000 UTC m=+209.376799844" observedRunningTime="2025-11-25 09:51:54.37006454 +0000 UTC m=+211.054020987" watchObservedRunningTime="2025-11-25 09:51:54.371442308 +0000 UTC m=+211.055398735" Nov 25 09:51:54 crc kubenswrapper[4926]: I1125 09:51:54.633354 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:51:54 crc kubenswrapper[4926]: I1125 09:51:54.633779 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:51:54 crc kubenswrapper[4926]: I1125 09:51:54.677782 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:51:55 crc kubenswrapper[4926]: I1125 09:51:55.363153 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerStarted","Data":"bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039"} Nov 25 09:51:55 crc kubenswrapper[4926]: I1125 09:51:55.467468 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:51:55 crc kubenswrapper[4926]: I1125 09:51:55.485793 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-brdt4" podStartSLOduration=3.760377793 podStartE2EDuration="1m3.485769011s" podCreationTimestamp="2025-11-25 09:50:52 +0000 UTC" firstStartedPulling="2025-11-25 09:50:54.699904356 +0000 UTC m=+151.383860783" lastFinishedPulling="2025-11-25 09:51:54.425295574 +0000 UTC m=+211.109252001" observedRunningTime="2025-11-25 09:51:55.387422275 +0000 UTC m=+212.071378702" watchObservedRunningTime="2025-11-25 09:51:55.485769011 +0000 UTC m=+212.169725438" Nov 25 09:51:55 crc kubenswrapper[4926]: I1125 09:51:55.836109 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:51:55 crc kubenswrapper[4926]: I1125 09:51:55.836403 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:51:55 crc kubenswrapper[4926]: I1125 09:51:55.876190 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:51:56 crc kubenswrapper[4926]: I1125 09:51:56.371146 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerStarted","Data":"ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31"} Nov 25 09:51:56 crc kubenswrapper[4926]: I1125 09:51:56.409347 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:51:57 crc kubenswrapper[4926]: I1125 09:51:57.406774 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n4xnt" podStartSLOduration=4.357706907 podStartE2EDuration="1m5.406750906s" podCreationTimestamp="2025-11-25 09:50:52 +0000 UTC" firstStartedPulling="2025-11-25 09:50:54.731403519 +0000 UTC m=+151.415359946" lastFinishedPulling="2025-11-25 09:51:55.780447518 +0000 UTC m=+212.464403945" observedRunningTime="2025-11-25 09:51:57.403121356 +0000 UTC m=+214.087077783" watchObservedRunningTime="2025-11-25 09:51:57.406750906 +0000 UTC m=+214.090707333" Nov 25 09:51:57 crc kubenswrapper[4926]: I1125 09:51:57.872008 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8bvj"] Nov 25 09:51:57 crc kubenswrapper[4926]: I1125 09:51:57.872303 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b8bvj" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="registry-server" containerID="cri-o://1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037" gracePeriod=2 Nov 25 09:51:58 crc kubenswrapper[4926]: I1125 09:51:58.395059 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerStarted","Data":"9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8"} Nov 25 09:51:58 crc kubenswrapper[4926]: I1125 09:51:58.417606 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2fxvt" podStartSLOduration=2.631900558 podStartE2EDuration="1m5.417583423s" podCreationTimestamp="2025-11-25 09:50:53 +0000 UTC" firstStartedPulling="2025-11-25 09:50:54.694807812 +0000 UTC m=+151.378764239" lastFinishedPulling="2025-11-25 09:51:57.480490687 +0000 UTC m=+214.164447104" observedRunningTime="2025-11-25 09:51:58.413336487 +0000 UTC m=+215.097292914" watchObservedRunningTime="2025-11-25 09:51:58.417583423 +0000 UTC m=+215.101539860" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.087360 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.233895 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-catalog-content\") pod \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.234018 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-utilities\") pod \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.234051 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zprk\" (UniqueName: \"kubernetes.io/projected/39af42eb-1c94-4a21-87bb-a370e45a3f1d-kube-api-access-4zprk\") pod \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\" (UID: \"39af42eb-1c94-4a21-87bb-a370e45a3f1d\") " Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.235334 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-utilities" (OuterVolumeSpecName: "utilities") pod "39af42eb-1c94-4a21-87bb-a370e45a3f1d" (UID: "39af42eb-1c94-4a21-87bb-a370e45a3f1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.246120 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39af42eb-1c94-4a21-87bb-a370e45a3f1d-kube-api-access-4zprk" (OuterVolumeSpecName: "kube-api-access-4zprk") pod "39af42eb-1c94-4a21-87bb-a370e45a3f1d" (UID: "39af42eb-1c94-4a21-87bb-a370e45a3f1d"). InnerVolumeSpecName "kube-api-access-4zprk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.255329 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39af42eb-1c94-4a21-87bb-a370e45a3f1d" (UID: "39af42eb-1c94-4a21-87bb-a370e45a3f1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.335967 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.336022 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39af42eb-1c94-4a21-87bb-a370e45a3f1d-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.336035 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zprk\" (UniqueName: \"kubernetes.io/projected/39af42eb-1c94-4a21-87bb-a370e45a3f1d-kube-api-access-4zprk\") on node \"crc\" DevicePath \"\"" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.404875 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs6rc" event={"ID":"d82538b8-f059-4fc6-bd88-622804100131","Type":"ContainerStarted","Data":"c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699"} Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.406919 4926 generic.go:334] "Generic (PLEG): container finished" podID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerID="1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037" exitCode=0 Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.406970 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8bvj" event={"ID":"39af42eb-1c94-4a21-87bb-a370e45a3f1d","Type":"ContainerDied","Data":"1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037"} Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.406973 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b8bvj" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.406989 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b8bvj" event={"ID":"39af42eb-1c94-4a21-87bb-a370e45a3f1d","Type":"ContainerDied","Data":"7d0fc311be6746642bf6f05a207ba2c9a8679c2517d8705dc583b71e16deef8c"} Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.407006 4926 scope.go:117] "RemoveContainer" containerID="1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.424321 4926 scope.go:117] "RemoveContainer" containerID="51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.433634 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qs6rc" podStartSLOduration=2.337521998 podStartE2EDuration="1m4.433602671s" podCreationTimestamp="2025-11-25 09:50:55 +0000 UTC" firstStartedPulling="2025-11-25 09:50:56.834908792 +0000 UTC m=+153.518865219" lastFinishedPulling="2025-11-25 09:51:58.930989465 +0000 UTC m=+215.614945892" observedRunningTime="2025-11-25 09:51:59.427848964 +0000 UTC m=+216.111805391" watchObservedRunningTime="2025-11-25 09:51:59.433602671 +0000 UTC m=+216.117559108" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.448696 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8bvj"] Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.458061 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b8bvj"] Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.459882 4926 scope.go:117] "RemoveContainer" containerID="d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.476241 4926 scope.go:117] "RemoveContainer" containerID="1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037" Nov 25 09:51:59 crc kubenswrapper[4926]: E1125 09:51:59.476799 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037\": container with ID starting with 1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037 not found: ID does not exist" containerID="1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.476857 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037"} err="failed to get container status \"1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037\": rpc error: code = NotFound desc = could not find container \"1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037\": container with ID starting with 1af8aa2029978684997416ad37b5789f58d3f5be7bd0d3cf0d66f6d7ba4e0037 not found: ID does not exist" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.476927 4926 scope.go:117] "RemoveContainer" containerID="51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd" Nov 25 09:51:59 crc kubenswrapper[4926]: E1125 09:51:59.477974 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd\": container with ID starting with 51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd not found: ID does not exist" containerID="51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.478008 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd"} err="failed to get container status \"51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd\": rpc error: code = NotFound desc = could not find container \"51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd\": container with ID starting with 51a9693a0bbfcd046652bc699284dcae33aac01d707c72865e9cb2dd5a8224cd not found: ID does not exist" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.478030 4926 scope.go:117] "RemoveContainer" containerID="d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28" Nov 25 09:51:59 crc kubenswrapper[4926]: E1125 09:51:59.478475 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28\": container with ID starting with d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28 not found: ID does not exist" containerID="d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28" Nov 25 09:51:59 crc kubenswrapper[4926]: I1125 09:51:59.478510 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28"} err="failed to get container status \"d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28\": rpc error: code = NotFound desc = could not find container \"d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28\": container with ID starting with d5c4fd0991810fec0b817eb4b88adb54ac4b777947c82d7440d01a0efd8c7a28 not found: ID does not exist" Nov 25 09:52:00 crc kubenswrapper[4926]: I1125 09:52:00.028524 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" path="/var/lib/kubelet/pods/39af42eb-1c94-4a21-87bb-a370e45a3f1d/volumes" Nov 25 09:52:00 crc kubenswrapper[4926]: I1125 09:52:00.271693 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jnhlw"] Nov 25 09:52:00 crc kubenswrapper[4926]: I1125 09:52:00.272009 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jnhlw" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="registry-server" containerID="cri-o://14169a6f0bb545959615df9c1036eaa2e37d5fc0e4647029bc26c5b82d2b1370" gracePeriod=2 Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.421251 4926 generic.go:334] "Generic (PLEG): container finished" podID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerID="14169a6f0bb545959615df9c1036eaa2e37d5fc0e4647029bc26c5b82d2b1370" exitCode=0 Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.421316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnhlw" event={"ID":"3b6717a6-c510-4800-bc0b-bf4506132cef","Type":"ContainerDied","Data":"14169a6f0bb545959615df9c1036eaa2e37d5fc0e4647029bc26c5b82d2b1370"} Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.797230 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.877714 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-utilities\") pod \"3b6717a6-c510-4800-bc0b-bf4506132cef\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.877853 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-catalog-content\") pod \"3b6717a6-c510-4800-bc0b-bf4506132cef\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.877974 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gft47\" (UniqueName: \"kubernetes.io/projected/3b6717a6-c510-4800-bc0b-bf4506132cef-kube-api-access-gft47\") pod \"3b6717a6-c510-4800-bc0b-bf4506132cef\" (UID: \"3b6717a6-c510-4800-bc0b-bf4506132cef\") " Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.878619 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-utilities" (OuterVolumeSpecName: "utilities") pod "3b6717a6-c510-4800-bc0b-bf4506132cef" (UID: "3b6717a6-c510-4800-bc0b-bf4506132cef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.886230 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b6717a6-c510-4800-bc0b-bf4506132cef-kube-api-access-gft47" (OuterVolumeSpecName: "kube-api-access-gft47") pod "3b6717a6-c510-4800-bc0b-bf4506132cef" (UID: "3b6717a6-c510-4800-bc0b-bf4506132cef"). InnerVolumeSpecName "kube-api-access-gft47". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.977105 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3b6717a6-c510-4800-bc0b-bf4506132cef" (UID: "3b6717a6-c510-4800-bc0b-bf4506132cef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.979795 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.979831 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gft47\" (UniqueName: \"kubernetes.io/projected/3b6717a6-c510-4800-bc0b-bf4506132cef-kube-api-access-gft47\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:01 crc kubenswrapper[4926]: I1125 09:52:01.979846 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3b6717a6-c510-4800-bc0b-bf4506132cef-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.387046 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.428079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jnhlw" event={"ID":"3b6717a6-c510-4800-bc0b-bf4506132cef","Type":"ContainerDied","Data":"51ad92ddd9840ab5e1fde4f49f4fc795239dca09022f5064a041da1df87f06af"} Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.428149 4926 scope.go:117] "RemoveContainer" containerID="14169a6f0bb545959615df9c1036eaa2e37d5fc0e4647029bc26c5b82d2b1370" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.428151 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jnhlw" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.441686 4926 scope.go:117] "RemoveContainer" containerID="da57a4320af15283c133274329225d2233e1191fe473be2cf6f92c81451e0410" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.444527 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.444583 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.451531 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jnhlw"] Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.455009 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jnhlw"] Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.461234 4926 scope.go:117] "RemoveContainer" containerID="16c3546f860b1a69781a75d691199e14a76dbc859c049baa934df377b91763c0" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.505289 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.669648 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.669747 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.709768 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.856811 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.856893 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.896034 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:52:02 crc kubenswrapper[4926]: I1125 09:52:02.905902 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerName="oauth-openshift" containerID="cri-o://e8aa38d0c206c2096a42aa21dd8e0c4518ae96f2f07589dda3be3e0eb0f7fa86" gracePeriod=15 Nov 25 09:52:03 crc kubenswrapper[4926]: I1125 09:52:03.437018 4926 generic.go:334] "Generic (PLEG): container finished" podID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerID="e8aa38d0c206c2096a42aa21dd8e0c4518ae96f2f07589dda3be3e0eb0f7fa86" exitCode=0 Nov 25 09:52:03 crc kubenswrapper[4926]: I1125 09:52:03.437867 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" event={"ID":"7bc1de70-8871-4225-8a3e-d6f849fc864a","Type":"ContainerDied","Data":"e8aa38d0c206c2096a42aa21dd8e0c4518ae96f2f07589dda3be3e0eb0f7fa86"} Nov 25 09:52:03 crc kubenswrapper[4926]: I1125 09:52:03.474761 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:52:03 crc kubenswrapper[4926]: I1125 09:52:03.488071 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:52:03 crc kubenswrapper[4926]: I1125 09:52:03.489146 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.021303 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" path="/var/lib/kubelet/pods/3b6717a6-c510-4800-bc0b-bf4506132cef/volumes" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.114835 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.114967 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.168882 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.293174 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.413395 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8668\" (UniqueName: \"kubernetes.io/projected/7bc1de70-8871-4225-8a3e-d6f849fc864a-kube-api-access-c8668\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.413464 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-dir\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.413508 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-idp-0-file-data\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.413562 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-ocp-branding-template\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.413599 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-cliconfig\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.413592 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.414453 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.414922 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-provider-selection\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.414985 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-trusted-ca-bundle\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415025 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-service-ca\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415071 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-session\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415111 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-error\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415140 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-serving-cert\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415172 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-login\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415195 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-router-certs\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415227 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-policies\") pod \"7bc1de70-8871-4225-8a3e-d6f849fc864a\" (UID: \"7bc1de70-8871-4225-8a3e-d6f849fc864a\") " Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415506 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415659 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.416175 4926 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.416199 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.415961 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.416686 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.419368 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.419414 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.419648 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.419834 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bc1de70-8871-4225-8a3e-d6f849fc864a-kube-api-access-c8668" (OuterVolumeSpecName: "kube-api-access-c8668") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "kube-api-access-c8668". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.419864 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.420541 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.420890 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.421054 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.421166 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "7bc1de70-8871-4225-8a3e-d6f849fc864a" (UID: "7bc1de70-8871-4225-8a3e-d6f849fc864a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.445301 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" event={"ID":"7bc1de70-8871-4225-8a3e-d6f849fc864a","Type":"ContainerDied","Data":"a04401f0eb89dd3a9ada364e71ef03d1d267aa865b50088989ffa5087b8a2e12"} Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.445368 4926 scope.go:117] "RemoveContainer" containerID="e8aa38d0c206c2096a42aa21dd8e0c4518ae96f2f07589dda3be3e0eb0f7fa86" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.445431 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ht99z" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.481319 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht99z"] Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.485602 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ht99z"] Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.491293 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517655 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517703 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517720 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517735 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517745 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517756 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517769 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517779 4926 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/7bc1de70-8871-4225-8a3e-d6f849fc864a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517787 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8668\" (UniqueName: \"kubernetes.io/projected/7bc1de70-8871-4225-8a3e-d6f849fc864a-kube-api-access-c8668\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517796 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.517804 4926 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/7bc1de70-8871-4225-8a3e-d6f849fc864a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:04 crc kubenswrapper[4926]: I1125 09:52:04.671120 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-brdt4"] Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.273610 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n4xnt"] Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.412731 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.413132 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.454310 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.455420 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-brdt4" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="registry-server" containerID="cri-o://bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039" gracePeriod=2 Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.456023 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n4xnt" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="registry-server" containerID="cri-o://ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31" gracePeriod=2 Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.498307 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.881143 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:52:05 crc kubenswrapper[4926]: I1125 09:52:05.886500 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.020755 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" path="/var/lib/kubelet/pods/7bc1de70-8871-4225-8a3e-d6f849fc864a/volumes" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.036705 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhk6k\" (UniqueName: \"kubernetes.io/projected/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-kube-api-access-jhk6k\") pod \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.036778 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cntdj\" (UniqueName: \"kubernetes.io/projected/f78474ad-ae46-4070-9049-bfd94d411642-kube-api-access-cntdj\") pod \"f78474ad-ae46-4070-9049-bfd94d411642\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.036849 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-catalog-content\") pod \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.037041 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-utilities\") pod \"f78474ad-ae46-4070-9049-bfd94d411642\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.037144 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-catalog-content\") pod \"f78474ad-ae46-4070-9049-bfd94d411642\" (UID: \"f78474ad-ae46-4070-9049-bfd94d411642\") " Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.037185 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-utilities\") pod \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\" (UID: \"78f5ad10-72d9-4dff-b9c4-54b9bfe33958\") " Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.037922 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-utilities" (OuterVolumeSpecName: "utilities") pod "f78474ad-ae46-4070-9049-bfd94d411642" (UID: "f78474ad-ae46-4070-9049-bfd94d411642"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.038264 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-utilities" (OuterVolumeSpecName: "utilities") pod "78f5ad10-72d9-4dff-b9c4-54b9bfe33958" (UID: "78f5ad10-72d9-4dff-b9c4-54b9bfe33958"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.047353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-kube-api-access-jhk6k" (OuterVolumeSpecName: "kube-api-access-jhk6k") pod "78f5ad10-72d9-4dff-b9c4-54b9bfe33958" (UID: "78f5ad10-72d9-4dff-b9c4-54b9bfe33958"). InnerVolumeSpecName "kube-api-access-jhk6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.048196 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f78474ad-ae46-4070-9049-bfd94d411642-kube-api-access-cntdj" (OuterVolumeSpecName: "kube-api-access-cntdj") pod "f78474ad-ae46-4070-9049-bfd94d411642" (UID: "f78474ad-ae46-4070-9049-bfd94d411642"). InnerVolumeSpecName "kube-api-access-cntdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.095267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "78f5ad10-72d9-4dff-b9c4-54b9bfe33958" (UID: "78f5ad10-72d9-4dff-b9c4-54b9bfe33958"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.095630 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f78474ad-ae46-4070-9049-bfd94d411642" (UID: "f78474ad-ae46-4070-9049-bfd94d411642"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.140536 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.140620 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.140638 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhk6k\" (UniqueName: \"kubernetes.io/projected/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-kube-api-access-jhk6k\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.140651 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cntdj\" (UniqueName: \"kubernetes.io/projected/f78474ad-ae46-4070-9049-bfd94d411642-kube-api-access-cntdj\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.140663 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/78f5ad10-72d9-4dff-b9c4-54b9bfe33958-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.140675 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f78474ad-ae46-4070-9049-bfd94d411642-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.463568 4926 generic.go:334] "Generic (PLEG): container finished" podID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerID="bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039" exitCode=0 Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.463684 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerDied","Data":"bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039"} Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.463726 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-brdt4" event={"ID":"78f5ad10-72d9-4dff-b9c4-54b9bfe33958","Type":"ContainerDied","Data":"ba5bc90ff67b67f201fe37455c9fe32c8e20e882768b124c708f505c300fa329"} Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.463724 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-brdt4" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.463772 4926 scope.go:117] "RemoveContainer" containerID="bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.467086 4926 generic.go:334] "Generic (PLEG): container finished" podID="f78474ad-ae46-4070-9049-bfd94d411642" containerID="ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31" exitCode=0 Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.467909 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n4xnt" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.468022 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerDied","Data":"ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31"} Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.468095 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n4xnt" event={"ID":"f78474ad-ae46-4070-9049-bfd94d411642","Type":"ContainerDied","Data":"bbdba7e97b417d3b64f9a02d0ff407de603607cffcc82fcb731f68b75900a07f"} Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.479381 4926 scope.go:117] "RemoveContainer" containerID="41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.495359 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-brdt4"] Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.498121 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-brdt4"] Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.508848 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n4xnt"] Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.512389 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n4xnt"] Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.515469 4926 scope.go:117] "RemoveContainer" containerID="c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.530620 4926 scope.go:117] "RemoveContainer" containerID="bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.531442 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039\": container with ID starting with bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039 not found: ID does not exist" containerID="bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.531505 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039"} err="failed to get container status \"bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039\": rpc error: code = NotFound desc = could not find container \"bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039\": container with ID starting with bd8e7b9f87e8278303a2fd40f0dd27846aa2a2e659986cbb0001c72c00142039 not found: ID does not exist" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.531546 4926 scope.go:117] "RemoveContainer" containerID="41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.532049 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2\": container with ID starting with 41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2 not found: ID does not exist" containerID="41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.532084 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2"} err="failed to get container status \"41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2\": rpc error: code = NotFound desc = could not find container \"41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2\": container with ID starting with 41961c4bf38ab867399dac57fb6ffa5bebbc932906c80705f41a0f88524948c2 not found: ID does not exist" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.532112 4926 scope.go:117] "RemoveContainer" containerID="c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.532809 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860\": container with ID starting with c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860 not found: ID does not exist" containerID="c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.532881 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860"} err="failed to get container status \"c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860\": rpc error: code = NotFound desc = could not find container \"c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860\": container with ID starting with c8c23a4fc82f85a1a3b5bc1760dbff30f9db5869697a9f46da77083b9f379860 not found: ID does not exist" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.532923 4926 scope.go:117] "RemoveContainer" containerID="ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.547232 4926 scope.go:117] "RemoveContainer" containerID="b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.564668 4926 scope.go:117] "RemoveContainer" containerID="7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.580852 4926 scope.go:117] "RemoveContainer" containerID="ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.582243 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31\": container with ID starting with ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31 not found: ID does not exist" containerID="ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.582388 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31"} err="failed to get container status \"ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31\": rpc error: code = NotFound desc = could not find container \"ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31\": container with ID starting with ce9d8533aca4070738f6ce3effd4c89e3dcd9054500623b9964dd9026cd63b31 not found: ID does not exist" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.582434 4926 scope.go:117] "RemoveContainer" containerID="b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.582850 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8\": container with ID starting with b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8 not found: ID does not exist" containerID="b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.582907 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8"} err="failed to get container status \"b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8\": rpc error: code = NotFound desc = could not find container \"b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8\": container with ID starting with b2957b33d7022896fe21a24cdb5e1833df5f495f62bc4d4cbbca378ebb4b54b8 not found: ID does not exist" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.582927 4926 scope.go:117] "RemoveContainer" containerID="7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.583243 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4\": container with ID starting with 7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4 not found: ID does not exist" containerID="7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.583279 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4"} err="failed to get container status \"7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4\": rpc error: code = NotFound desc = could not find container \"7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4\": container with ID starting with 7b6b342674f7ea24d695f00aef2921534530de33507725d53e38f6cdfcc321f4 not found: ID does not exist" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.601632 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-8476cd6899-s44p2"] Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.601975 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224" containerName="pruner" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.601992 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224" containerName="pruner" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602007 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602014 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602025 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602031 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602040 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602046 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602058 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd0f7444-df82-40cd-aaa3-60250afa1f36" containerName="collect-profiles" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602064 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd0f7444-df82-40cd-aaa3-60250afa1f36" containerName="collect-profiles" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602077 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602085 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602097 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602107 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602117 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerName="oauth-openshift" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602124 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerName="oauth-openshift" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602134 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602142 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="extract-content" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602152 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0414cbf3-5dca-4fb0-8e0c-093b7f9026ca" containerName="pruner" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602196 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0414cbf3-5dca-4fb0-8e0c-093b7f9026ca" containerName="pruner" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602273 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602283 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602291 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602297 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602305 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602311 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602322 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602328 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602334 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602341 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: E1125 09:52:06.602349 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602355 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="extract-utilities" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602477 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9aca11a8-1b90-4f9e-ad6e-fb4aeae0f224" containerName="pruner" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602486 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0414cbf3-5dca-4fb0-8e0c-093b7f9026ca" containerName="pruner" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602498 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bc1de70-8871-4225-8a3e-d6f849fc864a" containerName="oauth-openshift" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602507 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f78474ad-ae46-4070-9049-bfd94d411642" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602519 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="39af42eb-1c94-4a21-87bb-a370e45a3f1d" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602528 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd0f7444-df82-40cd-aaa3-60250afa1f36" containerName="collect-profiles" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602536 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b6717a6-c510-4800-bc0b-bf4506132cef" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.602544 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" containerName="registry-server" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.603058 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.612829 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.613071 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.613245 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.613609 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.613900 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.614136 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.612857 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.616184 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.616370 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.616590 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.616739 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.616798 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.621065 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.622659 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.626353 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8476cd6899-s44p2"] Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.627228 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747636 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-router-certs\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747692 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747760 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747817 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qjwc\" (UniqueName: \"kubernetes.io/projected/92b2b88b-3d77-46c9-aead-a635ad692e80-kube-api-access-2qjwc\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747849 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92b2b88b-3d77-46c9-aead-a635ad692e80-audit-dir\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747874 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-audit-policies\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.747960 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-login\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.748085 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.748238 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.748311 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-service-ca\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.748353 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.748446 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-session\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.748517 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-error\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849650 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-error\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849714 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-router-certs\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849774 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qjwc\" (UniqueName: \"kubernetes.io/projected/92b2b88b-3d77-46c9-aead-a635ad692e80-kube-api-access-2qjwc\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92b2b88b-3d77-46c9-aead-a635ad692e80-audit-dir\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849860 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-audit-policies\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849892 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.849928 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-login\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.851306 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.851711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.851750 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-service-ca\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.851784 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.851243 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-audit-policies\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.851830 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-session\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.850065 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/92b2b88b-3d77-46c9-aead-a635ad692e80-audit-dir\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.852596 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-cliconfig\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.852596 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-service-ca\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.852678 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.856291 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.856344 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-error\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.857515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-serving-cert\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.859387 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-session\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.859519 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.859780 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.865412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-user-template-login\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.865431 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/92b2b88b-3d77-46c9-aead-a635ad692e80-v4-0-config-system-router-certs\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.867473 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qjwc\" (UniqueName: \"kubernetes.io/projected/92b2b88b-3d77-46c9-aead-a635ad692e80-kube-api-access-2qjwc\") pod \"oauth-openshift-8476cd6899-s44p2\" (UID: \"92b2b88b-3d77-46c9-aead-a635ad692e80\") " pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:06 crc kubenswrapper[4926]: I1125 09:52:06.944987 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:07 crc kubenswrapper[4926]: I1125 09:52:07.124384 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-8476cd6899-s44p2"] Nov 25 09:52:07 crc kubenswrapper[4926]: I1125 09:52:07.477376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" event={"ID":"92b2b88b-3d77-46c9-aead-a635ad692e80","Type":"ContainerStarted","Data":"00e6df422bfc9e9350486d7cb3b248fde77817cc5fbd29104c4b45789ae7580f"} Nov 25 09:52:07 crc kubenswrapper[4926]: I1125 09:52:07.477448 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" event={"ID":"92b2b88b-3d77-46c9-aead-a635ad692e80","Type":"ContainerStarted","Data":"f5e7def521dcb6932e634d5e8fac9422f241f894f3a74c17f66ed8325808c431"} Nov 25 09:52:07 crc kubenswrapper[4926]: I1125 09:52:07.514367 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" podStartSLOduration=30.514339985 podStartE2EDuration="30.514339985s" podCreationTimestamp="2025-11-25 09:51:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:52:07.504436864 +0000 UTC m=+224.188393291" watchObservedRunningTime="2025-11-25 09:52:07.514339985 +0000 UTC m=+224.198296412" Nov 25 09:52:08 crc kubenswrapper[4926]: I1125 09:52:08.021641 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78f5ad10-72d9-4dff-b9c4-54b9bfe33958" path="/var/lib/kubelet/pods/78f5ad10-72d9-4dff-b9c4-54b9bfe33958/volumes" Nov 25 09:52:08 crc kubenswrapper[4926]: I1125 09:52:08.022635 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f78474ad-ae46-4070-9049-bfd94d411642" path="/var/lib/kubelet/pods/f78474ad-ae46-4070-9049-bfd94d411642/volumes" Nov 25 09:52:08 crc kubenswrapper[4926]: I1125 09:52:08.484337 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:08 crc kubenswrapper[4926]: I1125 09:52:08.489523 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-8476cd6899-s44p2" Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.015637 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.016004 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.019630 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.020337 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.020418 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d" gracePeriod=600 Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.522142 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d" exitCode=0 Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.522250 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d"} Nov 25 09:52:14 crc kubenswrapper[4926]: I1125 09:52:14.522489 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"2e7ab6176895234db45534a987817a5aad23e28ab3b2e4096d271630a3d44b38"} Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.883784 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5nkct"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.884738 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5nkct" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="registry-server" containerID="cri-o://50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba" gracePeriod=30 Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.889987 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tlltp"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.892034 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tlltp" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="registry-server" containerID="cri-o://63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691" gracePeriod=30 Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.907244 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nn4ms"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.907547 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerName="marketplace-operator" containerID="cri-o://1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46" gracePeriod=30 Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.926049 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tkd8b"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.927194 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.927623 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2fxvt"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.927918 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2fxvt" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="registry-server" containerID="cri-o://9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8" gracePeriod=30 Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.937290 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tkd8b"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.940232 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qs6rc"] Nov 25 09:52:40 crc kubenswrapper[4926]: I1125 09:52:40.940550 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qs6rc" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="registry-server" containerID="cri-o://c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699" gracePeriod=30 Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.096857 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lddc6\" (UniqueName: \"kubernetes.io/projected/8e239ba8-a966-470d-b462-c63630f38b7f-kube-api-access-lddc6\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.097122 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e239ba8-a966-470d-b462-c63630f38b7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.097359 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8e239ba8-a966-470d-b462-c63630f38b7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.199199 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e239ba8-a966-470d-b462-c63630f38b7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.199273 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8e239ba8-a966-470d-b462-c63630f38b7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.199309 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lddc6\" (UniqueName: \"kubernetes.io/projected/8e239ba8-a966-470d-b462-c63630f38b7f-kube-api-access-lddc6\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.202151 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8e239ba8-a966-470d-b462-c63630f38b7f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.208874 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/8e239ba8-a966-470d-b462-c63630f38b7f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.218047 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lddc6\" (UniqueName: \"kubernetes.io/projected/8e239ba8-a966-470d-b462-c63630f38b7f-kube-api-access-lddc6\") pod \"marketplace-operator-79b997595-tkd8b\" (UID: \"8e239ba8-a966-470d-b462-c63630f38b7f\") " pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.350851 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.367737 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.400890 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-utilities\") pod \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.400968 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-catalog-content\") pod \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.401001 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h99jg\" (UniqueName: \"kubernetes.io/projected/85c11326-80b9-4be7-beda-d9ef3be2cd8a-kube-api-access-h99jg\") pod \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\" (UID: \"85c11326-80b9-4be7-beda-d9ef3be2cd8a\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.401749 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-utilities" (OuterVolumeSpecName: "utilities") pod "85c11326-80b9-4be7-beda-d9ef3be2cd8a" (UID: "85c11326-80b9-4be7-beda-d9ef3be2cd8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.402686 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.403094 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.405911 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85c11326-80b9-4be7-beda-d9ef3be2cd8a-kube-api-access-h99jg" (OuterVolumeSpecName: "kube-api-access-h99jg") pod "85c11326-80b9-4be7-beda-d9ef3be2cd8a" (UID: "85c11326-80b9-4be7-beda-d9ef3be2cd8a"). InnerVolumeSpecName "kube-api-access-h99jg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.414392 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.454566 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.464962 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "85c11326-80b9-4be7-beda-d9ef3be2cd8a" (UID: "85c11326-80b9-4be7-beda-d9ef3be2cd8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502452 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-utilities\") pod \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502794 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-trusted-ca\") pod \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502820 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c84n7\" (UniqueName: \"kubernetes.io/projected/196347d9-90c7-41e5-a8c5-75acad8f047a-kube-api-access-c84n7\") pod \"196347d9-90c7-41e5-a8c5-75acad8f047a\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502848 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-catalog-content\") pod \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502904 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-operator-metrics\") pod \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502947 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r99kv\" (UniqueName: \"kubernetes.io/projected/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-kube-api-access-r99kv\") pod \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\" (UID: \"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.502967 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-utilities\") pod \"196347d9-90c7-41e5-a8c5-75acad8f047a\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.503004 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-utilities\") pod \"d82538b8-f059-4fc6-bd88-622804100131\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.503047 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-catalog-content\") pod \"196347d9-90c7-41e5-a8c5-75acad8f047a\" (UID: \"196347d9-90c7-41e5-a8c5-75acad8f047a\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.503131 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-925n5\" (UniqueName: \"kubernetes.io/projected/121f5347-8bf8-4d6c-ab81-3f22c87275bc-kube-api-access-925n5\") pod \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\" (UID: \"121f5347-8bf8-4d6c-ab81-3f22c87275bc\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.503165 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-catalog-content\") pod \"d82538b8-f059-4fc6-bd88-622804100131\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.503208 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgtcx\" (UniqueName: \"kubernetes.io/projected/d82538b8-f059-4fc6-bd88-622804100131-kube-api-access-hgtcx\") pod \"d82538b8-f059-4fc6-bd88-622804100131\" (UID: \"d82538b8-f059-4fc6-bd88-622804100131\") " Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.503584 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-utilities" (OuterVolumeSpecName: "utilities") pod "7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" (UID: "7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.504607 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.504630 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/85c11326-80b9-4be7-beda-d9ef3be2cd8a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.504647 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h99jg\" (UniqueName: \"kubernetes.io/projected/85c11326-80b9-4be7-beda-d9ef3be2cd8a-kube-api-access-h99jg\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.504660 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.505876 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "121f5347-8bf8-4d6c-ab81-3f22c87275bc" (UID: "121f5347-8bf8-4d6c-ab81-3f22c87275bc"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.506046 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-utilities" (OuterVolumeSpecName: "utilities") pod "196347d9-90c7-41e5-a8c5-75acad8f047a" (UID: "196347d9-90c7-41e5-a8c5-75acad8f047a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.507484 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d82538b8-f059-4fc6-bd88-622804100131-kube-api-access-hgtcx" (OuterVolumeSpecName: "kube-api-access-hgtcx") pod "d82538b8-f059-4fc6-bd88-622804100131" (UID: "d82538b8-f059-4fc6-bd88-622804100131"). InnerVolumeSpecName "kube-api-access-hgtcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.507585 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-utilities" (OuterVolumeSpecName: "utilities") pod "d82538b8-f059-4fc6-bd88-622804100131" (UID: "d82538b8-f059-4fc6-bd88-622804100131"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.507614 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/196347d9-90c7-41e5-a8c5-75acad8f047a-kube-api-access-c84n7" (OuterVolumeSpecName: "kube-api-access-c84n7") pod "196347d9-90c7-41e5-a8c5-75acad8f047a" (UID: "196347d9-90c7-41e5-a8c5-75acad8f047a"). InnerVolumeSpecName "kube-api-access-c84n7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.509270 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/121f5347-8bf8-4d6c-ab81-3f22c87275bc-kube-api-access-925n5" (OuterVolumeSpecName: "kube-api-access-925n5") pod "121f5347-8bf8-4d6c-ab81-3f22c87275bc" (UID: "121f5347-8bf8-4d6c-ab81-3f22c87275bc"). InnerVolumeSpecName "kube-api-access-925n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.510146 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-kube-api-access-r99kv" (OuterVolumeSpecName: "kube-api-access-r99kv") pod "7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" (UID: "7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c"). InnerVolumeSpecName "kube-api-access-r99kv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.527769 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "196347d9-90c7-41e5-a8c5-75acad8f047a" (UID: "196347d9-90c7-41e5-a8c5-75acad8f047a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.539044 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "121f5347-8bf8-4d6c-ab81-3f22c87275bc" (UID: "121f5347-8bf8-4d6c-ab81-3f22c87275bc"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.589861 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" (UID: "7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606122 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606165 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-925n5\" (UniqueName: \"kubernetes.io/projected/121f5347-8bf8-4d6c-ab81-3f22c87275bc-kube-api-access-925n5\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606182 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgtcx\" (UniqueName: \"kubernetes.io/projected/d82538b8-f059-4fc6-bd88-622804100131-kube-api-access-hgtcx\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606195 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606206 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c84n7\" (UniqueName: \"kubernetes.io/projected/196347d9-90c7-41e5-a8c5-75acad8f047a-kube-api-access-c84n7\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606215 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606224 4926 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/121f5347-8bf8-4d6c-ab81-3f22c87275bc-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606233 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r99kv\" (UniqueName: \"kubernetes.io/projected/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c-kube-api-access-r99kv\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606242 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/196347d9-90c7-41e5-a8c5-75acad8f047a-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.606251 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.633170 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d82538b8-f059-4fc6-bd88-622804100131" (UID: "d82538b8-f059-4fc6-bd88-622804100131"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.670777 4926 generic.go:334] "Generic (PLEG): container finished" podID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerID="1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46" exitCode=0 Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.670835 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" event={"ID":"121f5347-8bf8-4d6c-ab81-3f22c87275bc","Type":"ContainerDied","Data":"1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.670883 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" event={"ID":"121f5347-8bf8-4d6c-ab81-3f22c87275bc","Type":"ContainerDied","Data":"c0f58f67bd261415d14d9c7a8937b818c5270f075cbd9d4696be93b958c4e535"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.670882 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nn4ms" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.670954 4926 scope.go:117] "RemoveContainer" containerID="1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.674427 4926 generic.go:334] "Generic (PLEG): container finished" podID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerID="9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8" exitCode=0 Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.674503 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerDied","Data":"9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.674518 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2fxvt" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.674541 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2fxvt" event={"ID":"196347d9-90c7-41e5-a8c5-75acad8f047a","Type":"ContainerDied","Data":"51fe78f7135327e3d4884b7ad3ebd62ba77833cb1bc5f16eaee0b1ad7a32486c"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.676585 4926 generic.go:334] "Generic (PLEG): container finished" podID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerID="63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691" exitCode=0 Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.676714 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tlltp" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.677045 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerDied","Data":"63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.677082 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tlltp" event={"ID":"7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c","Type":"ContainerDied","Data":"c8ae29ea7608ef84a658a9ea87e7da600d1b899ae67a9e0ca2e888391a2d4b1f"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.684166 4926 generic.go:334] "Generic (PLEG): container finished" podID="d82538b8-f059-4fc6-bd88-622804100131" containerID="c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699" exitCode=0 Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.684504 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qs6rc" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.684881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs6rc" event={"ID":"d82538b8-f059-4fc6-bd88-622804100131","Type":"ContainerDied","Data":"c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.686179 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qs6rc" event={"ID":"d82538b8-f059-4fc6-bd88-622804100131","Type":"ContainerDied","Data":"28e5f3f020a1e3a411a49bde3fde59600c26a4202688235ff35d7729d277fcf8"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.689909 4926 generic.go:334] "Generic (PLEG): container finished" podID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerID="50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba" exitCode=0 Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.689958 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5nkct" event={"ID":"85c11326-80b9-4be7-beda-d9ef3be2cd8a","Type":"ContainerDied","Data":"50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.689981 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5nkct" event={"ID":"85c11326-80b9-4be7-beda-d9ef3be2cd8a","Type":"ContainerDied","Data":"b543467a386d2390b41600a897ec19e6dcf141f2e0d8d806f5e1fec6221d2ce1"} Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.690386 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5nkct" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.692691 4926 scope.go:117] "RemoveContainer" containerID="1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.694356 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46\": container with ID starting with 1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46 not found: ID does not exist" containerID="1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.695203 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46"} err="failed to get container status \"1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46\": rpc error: code = NotFound desc = could not find container \"1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46\": container with ID starting with 1bbf5f9a0b91f650e75aebecf3f5d9d24a814826b3b10ef20bf8bb1b26f9fc46 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.695289 4926 scope.go:117] "RemoveContainer" containerID="9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.708046 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d82538b8-f059-4fc6-bd88-622804100131-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.714061 4926 scope.go:117] "RemoveContainer" containerID="c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.720146 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nn4ms"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.721718 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nn4ms"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.747073 4926 scope.go:117] "RemoveContainer" containerID="0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.747264 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2fxvt"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.749206 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2fxvt"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.774316 4926 scope.go:117] "RemoveContainer" containerID="9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.774966 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8\": container with ID starting with 9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8 not found: ID does not exist" containerID="9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.775007 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8"} err="failed to get container status \"9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8\": rpc error: code = NotFound desc = could not find container \"9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8\": container with ID starting with 9161c83532d46c83024c96f83eebc4769aebb8655dd1f03c90773341bfcd0bc8 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.775036 4926 scope.go:117] "RemoveContainer" containerID="c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.775533 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10\": container with ID starting with c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10 not found: ID does not exist" containerID="c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.775556 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10"} err="failed to get container status \"c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10\": rpc error: code = NotFound desc = could not find container \"c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10\": container with ID starting with c727fffacb29ce603616748a156e044363247df75cae56454bacec48e7413d10 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.775571 4926 scope.go:117] "RemoveContainer" containerID="0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.775920 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8\": container with ID starting with 0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8 not found: ID does not exist" containerID="0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.775998 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8"} err="failed to get container status \"0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8\": rpc error: code = NotFound desc = could not find container \"0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8\": container with ID starting with 0ff4bf05d58d79487e00e8a2f85e0ba4a526e8e541863a32a6850585205e2fd8 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.776012 4926 scope.go:117] "RemoveContainer" containerID="63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.778318 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qs6rc"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.781340 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qs6rc"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.798327 4926 scope.go:117] "RemoveContainer" containerID="6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.802592 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5nkct"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.807232 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5nkct"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.811063 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tlltp"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.814208 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tlltp"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.816376 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-tkd8b"] Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.817190 4926 scope.go:117] "RemoveContainer" containerID="edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.831356 4926 scope.go:117] "RemoveContainer" containerID="63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.832017 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691\": container with ID starting with 63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691 not found: ID does not exist" containerID="63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.832059 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691"} err="failed to get container status \"63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691\": rpc error: code = NotFound desc = could not find container \"63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691\": container with ID starting with 63a545b90dab7d6ecb05365ca2a5018bfb2517bd53b79eabf99b6487b6352691 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.832089 4926 scope.go:117] "RemoveContainer" containerID="6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.834170 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346\": container with ID starting with 6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346 not found: ID does not exist" containerID="6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.834224 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346"} err="failed to get container status \"6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346\": rpc error: code = NotFound desc = could not find container \"6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346\": container with ID starting with 6a3b97fd4808df463e12d40b2d7716e38a4cfc07ee074d4ee89137be42425346 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.834260 4926 scope.go:117] "RemoveContainer" containerID="edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.834727 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d\": container with ID starting with edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d not found: ID does not exist" containerID="edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.834769 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d"} err="failed to get container status \"edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d\": rpc error: code = NotFound desc = could not find container \"edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d\": container with ID starting with edee2f3386d319254f02f571d1004a146455d64bec06ebe01e7fc254f9c1ce3d not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.834802 4926 scope.go:117] "RemoveContainer" containerID="c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.850998 4926 scope.go:117] "RemoveContainer" containerID="116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.870817 4926 scope.go:117] "RemoveContainer" containerID="be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.888466 4926 scope.go:117] "RemoveContainer" containerID="c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.889649 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699\": container with ID starting with c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699 not found: ID does not exist" containerID="c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.889682 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699"} err="failed to get container status \"c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699\": rpc error: code = NotFound desc = could not find container \"c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699\": container with ID starting with c8803786a746ede56b5ac576f5155541866bb7a4af39bf117208680666f03699 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.889708 4926 scope.go:117] "RemoveContainer" containerID="116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.889977 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514\": container with ID starting with 116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514 not found: ID does not exist" containerID="116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.890002 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514"} err="failed to get container status \"116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514\": rpc error: code = NotFound desc = could not find container \"116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514\": container with ID starting with 116d7686cffafb85a8a2a05445b8c34db592a57852007a8d92a3216f21ab9514 not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.890020 4926 scope.go:117] "RemoveContainer" containerID="be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.890371 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa\": container with ID starting with be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa not found: ID does not exist" containerID="be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.890389 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa"} err="failed to get container status \"be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa\": rpc error: code = NotFound desc = could not find container \"be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa\": container with ID starting with be8d499fdb7884ef0e47fdbc86c112c87fa448fa44a63564cdb6ab9e6f2360fa not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.890404 4926 scope.go:117] "RemoveContainer" containerID="50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.920379 4926 scope.go:117] "RemoveContainer" containerID="8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.941435 4926 scope.go:117] "RemoveContainer" containerID="157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.959202 4926 scope.go:117] "RemoveContainer" containerID="50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.959835 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba\": container with ID starting with 50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba not found: ID does not exist" containerID="50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.959903 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba"} err="failed to get container status \"50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba\": rpc error: code = NotFound desc = could not find container \"50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba\": container with ID starting with 50e0269428c02cd6504fe645c73b54ca068efc0d7b3443dc0f4a534e8153e7ba not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.959971 4926 scope.go:117] "RemoveContainer" containerID="8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.962533 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d\": container with ID starting with 8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d not found: ID does not exist" containerID="8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.962576 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d"} err="failed to get container status \"8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d\": rpc error: code = NotFound desc = could not find container \"8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d\": container with ID starting with 8abee487ba5b8ece69c6342ad89c71f2fa86a160fb8a635ecd266f1b0b03b66d not found: ID does not exist" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.962604 4926 scope.go:117] "RemoveContainer" containerID="157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d" Nov 25 09:52:41 crc kubenswrapper[4926]: E1125 09:52:41.963166 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d\": container with ID starting with 157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d not found: ID does not exist" containerID="157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d" Nov 25 09:52:41 crc kubenswrapper[4926]: I1125 09:52:41.963214 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d"} err="failed to get container status \"157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d\": rpc error: code = NotFound desc = could not find container \"157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d\": container with ID starting with 157b59ef3e313856cf95bad6541a499f8cb3fe9ca595454f45d7947ed149821d not found: ID does not exist" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.020562 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" path="/var/lib/kubelet/pods/121f5347-8bf8-4d6c-ab81-3f22c87275bc/volumes" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.021209 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" path="/var/lib/kubelet/pods/196347d9-90c7-41e5-a8c5-75acad8f047a/volumes" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.021793 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" path="/var/lib/kubelet/pods/7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c/volumes" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.022771 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" path="/var/lib/kubelet/pods/85c11326-80b9-4be7-beda-d9ef3be2cd8a/volumes" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.023325 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d82538b8-f059-4fc6-bd88-622804100131" path="/var/lib/kubelet/pods/d82538b8-f059-4fc6-bd88-622804100131/volumes" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.702287 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" event={"ID":"8e239ba8-a966-470d-b462-c63630f38b7f","Type":"ContainerStarted","Data":"60e71f9e5eeccec9e5fd287702d4d8b475309c3b1d255aa09006d2cf42ed7c65"} Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.702346 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" event={"ID":"8e239ba8-a966-470d-b462-c63630f38b7f","Type":"ContainerStarted","Data":"9553f6fdfd73fbb48cadc067f28a3e2b63a5e55e3623565b5e45de86cef11197"} Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.702502 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.706284 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 09:52:42 crc kubenswrapper[4926]: I1125 09:52:42.721958 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" podStartSLOduration=2.721907617 podStartE2EDuration="2.721907617s" podCreationTimestamp="2025-11-25 09:52:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:52:42.719532242 +0000 UTC m=+259.403488669" watchObservedRunningTime="2025-11-25 09:52:42.721907617 +0000 UTC m=+259.405864064" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.095710 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zjvdf"] Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.095991 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096006 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096018 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096024 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096032 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096039 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096048 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096054 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096061 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerName="marketplace-operator" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096067 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerName="marketplace-operator" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096075 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096080 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096088 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096094 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096101 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096107 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096115 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096120 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096128 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096134 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096144 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096150 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="extract-utilities" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096159 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096165 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="extract-content" Nov 25 09:52:43 crc kubenswrapper[4926]: E1125 09:52:43.096175 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096181 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096261 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e2d1b08-b9b1-4097-9c40-f9d35aa85d6c" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096273 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d82538b8-f059-4fc6-bd88-622804100131" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096279 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="121f5347-8bf8-4d6c-ab81-3f22c87275bc" containerName="marketplace-operator" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096288 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="85c11326-80b9-4be7-beda-d9ef3be2cd8a" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.096299 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="196347d9-90c7-41e5-a8c5-75acad8f047a" containerName="registry-server" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.100200 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.102122 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zjvdf"] Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.102326 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.228592 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f0105d-9770-47c3-bfd1-02708d16cd5d-utilities\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.228662 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bprxj\" (UniqueName: \"kubernetes.io/projected/23f0105d-9770-47c3-bfd1-02708d16cd5d-kube-api-access-bprxj\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.228950 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f0105d-9770-47c3-bfd1-02708d16cd5d-catalog-content\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.298998 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-cbpg8"] Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.300559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.305559 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.308256 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbpg8"] Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330310 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/159d4219-a0ff-4fc1-a29c-63fa49502aac-utilities\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330377 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/159d4219-a0ff-4fc1-a29c-63fa49502aac-catalog-content\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330406 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f0105d-9770-47c3-bfd1-02708d16cd5d-utilities\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330425 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bprxj\" (UniqueName: \"kubernetes.io/projected/23f0105d-9770-47c3-bfd1-02708d16cd5d-kube-api-access-bprxj\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330448 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f0105d-9770-47c3-bfd1-02708d16cd5d-catalog-content\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330475 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkll4\" (UniqueName: \"kubernetes.io/projected/159d4219-a0ff-4fc1-a29c-63fa49502aac-kube-api-access-vkll4\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.330850 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23f0105d-9770-47c3-bfd1-02708d16cd5d-utilities\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.331027 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23f0105d-9770-47c3-bfd1-02708d16cd5d-catalog-content\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.351588 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bprxj\" (UniqueName: \"kubernetes.io/projected/23f0105d-9770-47c3-bfd1-02708d16cd5d-kube-api-access-bprxj\") pod \"certified-operators-zjvdf\" (UID: \"23f0105d-9770-47c3-bfd1-02708d16cd5d\") " pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.431773 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/159d4219-a0ff-4fc1-a29c-63fa49502aac-utilities\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.431841 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/159d4219-a0ff-4fc1-a29c-63fa49502aac-catalog-content\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.431875 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkll4\" (UniqueName: \"kubernetes.io/projected/159d4219-a0ff-4fc1-a29c-63fa49502aac-kube-api-access-vkll4\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.432583 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/159d4219-a0ff-4fc1-a29c-63fa49502aac-utilities\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.432683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/159d4219-a0ff-4fc1-a29c-63fa49502aac-catalog-content\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.436278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.449204 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkll4\" (UniqueName: \"kubernetes.io/projected/159d4219-a0ff-4fc1-a29c-63fa49502aac-kube-api-access-vkll4\") pod \"redhat-marketplace-cbpg8\" (UID: \"159d4219-a0ff-4fc1-a29c-63fa49502aac\") " pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.643308 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zjvdf"] Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.644580 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.753534 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjvdf" event={"ID":"23f0105d-9770-47c3-bfd1-02708d16cd5d","Type":"ContainerStarted","Data":"4b98f4f52da9b8b22cb17ba08297e5bc515db950f768511a11f4405f1c6b22e9"} Nov 25 09:52:43 crc kubenswrapper[4926]: I1125 09:52:43.834202 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-cbpg8"] Nov 25 09:52:43 crc kubenswrapper[4926]: W1125 09:52:43.914201 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod159d4219_a0ff_4fc1_a29c_63fa49502aac.slice/crio-68629685b0ec7a89582448b97b09cbd4475d4ae7d2d52de87c3edf0ae2875c28 WatchSource:0}: Error finding container 68629685b0ec7a89582448b97b09cbd4475d4ae7d2d52de87c3edf0ae2875c28: Status 404 returned error can't find the container with id 68629685b0ec7a89582448b97b09cbd4475d4ae7d2d52de87c3edf0ae2875c28 Nov 25 09:52:44 crc kubenswrapper[4926]: I1125 09:52:44.760122 4926 generic.go:334] "Generic (PLEG): container finished" podID="23f0105d-9770-47c3-bfd1-02708d16cd5d" containerID="5c5d9ace457bcdac56467b0d100124aba4fe608a137fd1256b13c46f580cf668" exitCode=0 Nov 25 09:52:44 crc kubenswrapper[4926]: I1125 09:52:44.760238 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjvdf" event={"ID":"23f0105d-9770-47c3-bfd1-02708d16cd5d","Type":"ContainerDied","Data":"5c5d9ace457bcdac56467b0d100124aba4fe608a137fd1256b13c46f580cf668"} Nov 25 09:52:44 crc kubenswrapper[4926]: I1125 09:52:44.762489 4926 generic.go:334] "Generic (PLEG): container finished" podID="159d4219-a0ff-4fc1-a29c-63fa49502aac" containerID="5370b937870e3d19f3fe529744c55a8b2d028daee094cdd925a6b1b778898e83" exitCode=0 Nov 25 09:52:44 crc kubenswrapper[4926]: I1125 09:52:44.763608 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbpg8" event={"ID":"159d4219-a0ff-4fc1-a29c-63fa49502aac","Type":"ContainerDied","Data":"5370b937870e3d19f3fe529744c55a8b2d028daee094cdd925a6b1b778898e83"} Nov 25 09:52:44 crc kubenswrapper[4926]: I1125 09:52:44.763645 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbpg8" event={"ID":"159d4219-a0ff-4fc1-a29c-63fa49502aac","Type":"ContainerStarted","Data":"68629685b0ec7a89582448b97b09cbd4475d4ae7d2d52de87c3edf0ae2875c28"} Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.498997 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-szd2l"] Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.501993 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.505271 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.510994 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-szd2l"] Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.570166 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kpl5\" (UniqueName: \"kubernetes.io/projected/6225b001-71a4-41f6-8122-b7abcbcfbe1c-kube-api-access-7kpl5\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.570307 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6225b001-71a4-41f6-8122-b7abcbcfbe1c-utilities\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.570350 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6225b001-71a4-41f6-8122-b7abcbcfbe1c-catalog-content\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.672698 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kpl5\" (UniqueName: \"kubernetes.io/projected/6225b001-71a4-41f6-8122-b7abcbcfbe1c-kube-api-access-7kpl5\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.673127 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6225b001-71a4-41f6-8122-b7abcbcfbe1c-utilities\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.673292 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6225b001-71a4-41f6-8122-b7abcbcfbe1c-catalog-content\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.673643 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6225b001-71a4-41f6-8122-b7abcbcfbe1c-utilities\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.673954 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6225b001-71a4-41f6-8122-b7abcbcfbe1c-catalog-content\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.700955 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kpl5\" (UniqueName: \"kubernetes.io/projected/6225b001-71a4-41f6-8122-b7abcbcfbe1c-kube-api-access-7kpl5\") pod \"community-operators-szd2l\" (UID: \"6225b001-71a4-41f6-8122-b7abcbcfbe1c\") " pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.708775 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-9r4r6"] Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.710867 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.713406 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.719083 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9r4r6"] Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.769729 4926 generic.go:334] "Generic (PLEG): container finished" podID="23f0105d-9770-47c3-bfd1-02708d16cd5d" containerID="554795655df0f60a60b2397ae15757c67f25173a37cdab69a74c09268e992e7a" exitCode=0 Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.769788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjvdf" event={"ID":"23f0105d-9770-47c3-bfd1-02708d16cd5d","Type":"ContainerDied","Data":"554795655df0f60a60b2397ae15757c67f25173a37cdab69a74c09268e992e7a"} Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.773623 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbpg8" event={"ID":"159d4219-a0ff-4fc1-a29c-63fa49502aac","Type":"ContainerDied","Data":"947d2ee2c3a392b9bf40840e1f5b9a9eeba6c7b66f268a638febf11354ef7efd"} Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.773615 4926 generic.go:334] "Generic (PLEG): container finished" podID="159d4219-a0ff-4fc1-a29c-63fa49502aac" containerID="947d2ee2c3a392b9bf40840e1f5b9a9eeba6c7b66f268a638febf11354ef7efd" exitCode=0 Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.775722 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-utilities\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.775779 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-catalog-content\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.775861 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b74zs\" (UniqueName: \"kubernetes.io/projected/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-kube-api-access-b74zs\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.862858 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.876907 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b74zs\" (UniqueName: \"kubernetes.io/projected/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-kube-api-access-b74zs\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.877004 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-utilities\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.877065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-catalog-content\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.877727 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-utilities\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.877863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-catalog-content\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:45 crc kubenswrapper[4926]: I1125 09:52:45.898451 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b74zs\" (UniqueName: \"kubernetes.io/projected/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-kube-api-access-b74zs\") pod \"redhat-operators-9r4r6\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.159664 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.268599 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-szd2l"] Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.607185 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-9r4r6"] Nov 25 09:52:46 crc kubenswrapper[4926]: W1125 09:52:46.610282 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod835a5c57_f564_4d91_a2fa_e8b7c8d576b3.slice/crio-eee673dbe13d9be2ba6590bbc82bccd78bc624730fdb6621a46b33cd82aeac84 WatchSource:0}: Error finding container eee673dbe13d9be2ba6590bbc82bccd78bc624730fdb6621a46b33cd82aeac84: Status 404 returned error can't find the container with id eee673dbe13d9be2ba6590bbc82bccd78bc624730fdb6621a46b33cd82aeac84 Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.783044 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zjvdf" event={"ID":"23f0105d-9770-47c3-bfd1-02708d16cd5d","Type":"ContainerStarted","Data":"17d1ecf1b103ebc3a1cd6173897a3548f23b36065155185fc2e16e2f2ceb84e3"} Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.785308 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-cbpg8" event={"ID":"159d4219-a0ff-4fc1-a29c-63fa49502aac","Type":"ContainerStarted","Data":"25a7dd64f483ba129a2ea2f7209b35000673e3c9889c442861a50fd6dbdefe70"} Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.790427 4926 generic.go:334] "Generic (PLEG): container finished" podID="6225b001-71a4-41f6-8122-b7abcbcfbe1c" containerID="f7a5e72d9332f62df0101313d347c8c7d36dd7e74d25a5a2cb78592200540ed8" exitCode=0 Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.790524 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szd2l" event={"ID":"6225b001-71a4-41f6-8122-b7abcbcfbe1c","Type":"ContainerDied","Data":"f7a5e72d9332f62df0101313d347c8c7d36dd7e74d25a5a2cb78592200540ed8"} Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.790549 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szd2l" event={"ID":"6225b001-71a4-41f6-8122-b7abcbcfbe1c","Type":"ContainerStarted","Data":"2a347dfc85293bb274af8304a601ad9aca55ea4bf26d1507d9b73425b6692484"} Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.795065 4926 generic.go:334] "Generic (PLEG): container finished" podID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerID="f5550879b016fd8d2f094d2a8bc8296d8d402bf6eb50c1cd4e86ef4ccd956791" exitCode=0 Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.795127 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerDied","Data":"f5550879b016fd8d2f094d2a8bc8296d8d402bf6eb50c1cd4e86ef4ccd956791"} Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.795169 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerStarted","Data":"eee673dbe13d9be2ba6590bbc82bccd78bc624730fdb6621a46b33cd82aeac84"} Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.823725 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zjvdf" podStartSLOduration=2.451728086 podStartE2EDuration="3.823700975s" podCreationTimestamp="2025-11-25 09:52:43 +0000 UTC" firstStartedPulling="2025-11-25 09:52:44.761819772 +0000 UTC m=+261.445776199" lastFinishedPulling="2025-11-25 09:52:46.133792661 +0000 UTC m=+262.817749088" observedRunningTime="2025-11-25 09:52:46.805198766 +0000 UTC m=+263.489155213" watchObservedRunningTime="2025-11-25 09:52:46.823700975 +0000 UTC m=+263.507657402" Nov 25 09:52:46 crc kubenswrapper[4926]: I1125 09:52:46.824322 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-cbpg8" podStartSLOduration=2.084122549 podStartE2EDuration="3.824315713s" podCreationTimestamp="2025-11-25 09:52:43 +0000 UTC" firstStartedPulling="2025-11-25 09:52:44.765320168 +0000 UTC m=+261.449276595" lastFinishedPulling="2025-11-25 09:52:46.505513332 +0000 UTC m=+263.189469759" observedRunningTime="2025-11-25 09:52:46.821061568 +0000 UTC m=+263.505017995" watchObservedRunningTime="2025-11-25 09:52:46.824315713 +0000 UTC m=+263.508272140" Nov 25 09:52:47 crc kubenswrapper[4926]: I1125 09:52:47.804426 4926 generic.go:334] "Generic (PLEG): container finished" podID="6225b001-71a4-41f6-8122-b7abcbcfbe1c" containerID="b3254c429a2430add72b8f7a191050b5772a533663df45973236fbdaca813848" exitCode=0 Nov 25 09:52:47 crc kubenswrapper[4926]: I1125 09:52:47.804538 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szd2l" event={"ID":"6225b001-71a4-41f6-8122-b7abcbcfbe1c","Type":"ContainerDied","Data":"b3254c429a2430add72b8f7a191050b5772a533663df45973236fbdaca813848"} Nov 25 09:52:47 crc kubenswrapper[4926]: I1125 09:52:47.807088 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerStarted","Data":"6bd9980837c606738969e00179eac73b25611bb64eca8a67d5193133f5597aa3"} Nov 25 09:52:48 crc kubenswrapper[4926]: I1125 09:52:48.817719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-szd2l" event={"ID":"6225b001-71a4-41f6-8122-b7abcbcfbe1c","Type":"ContainerStarted","Data":"2c36968f49ae0cb1344c5dc112c6b27a9e0dad8d2fc444e204952eb8b8d8d40d"} Nov 25 09:52:48 crc kubenswrapper[4926]: I1125 09:52:48.821910 4926 generic.go:334] "Generic (PLEG): container finished" podID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerID="6bd9980837c606738969e00179eac73b25611bb64eca8a67d5193133f5597aa3" exitCode=0 Nov 25 09:52:48 crc kubenswrapper[4926]: I1125 09:52:48.822021 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerDied","Data":"6bd9980837c606738969e00179eac73b25611bb64eca8a67d5193133f5597aa3"} Nov 25 09:52:48 crc kubenswrapper[4926]: I1125 09:52:48.859187 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-szd2l" podStartSLOduration=2.426120231 podStartE2EDuration="3.859164668s" podCreationTimestamp="2025-11-25 09:52:45 +0000 UTC" firstStartedPulling="2025-11-25 09:52:46.791802706 +0000 UTC m=+263.475759133" lastFinishedPulling="2025-11-25 09:52:48.224847143 +0000 UTC m=+264.908803570" observedRunningTime="2025-11-25 09:52:48.840109323 +0000 UTC m=+265.524065770" watchObservedRunningTime="2025-11-25 09:52:48.859164668 +0000 UTC m=+265.543121105" Nov 25 09:52:50 crc kubenswrapper[4926]: I1125 09:52:50.844612 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerStarted","Data":"3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96"} Nov 25 09:52:50 crc kubenswrapper[4926]: I1125 09:52:50.865462 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-9r4r6" podStartSLOduration=3.407187911 podStartE2EDuration="5.865439102s" podCreationTimestamp="2025-11-25 09:52:45 +0000 UTC" firstStartedPulling="2025-11-25 09:52:46.796474022 +0000 UTC m=+263.480430449" lastFinishedPulling="2025-11-25 09:52:49.254725213 +0000 UTC m=+265.938681640" observedRunningTime="2025-11-25 09:52:50.864320569 +0000 UTC m=+267.548276996" watchObservedRunningTime="2025-11-25 09:52:50.865439102 +0000 UTC m=+267.549395529" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.437087 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.438465 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.487339 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.634129 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.645766 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.677571 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.898828 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zjvdf" Nov 25 09:52:53 crc kubenswrapper[4926]: I1125 09:52:53.901308 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-cbpg8" Nov 25 09:52:55 crc kubenswrapper[4926]: I1125 09:52:55.863919 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:55 crc kubenswrapper[4926]: I1125 09:52:55.864272 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:55 crc kubenswrapper[4926]: I1125 09:52:55.902162 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:55 crc kubenswrapper[4926]: I1125 09:52:55.943823 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-szd2l" Nov 25 09:52:56 crc kubenswrapper[4926]: I1125 09:52:56.160002 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:56 crc kubenswrapper[4926]: I1125 09:52:56.160063 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:56 crc kubenswrapper[4926]: I1125 09:52:56.196422 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:52:56 crc kubenswrapper[4926]: I1125 09:52:56.939576 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 09:54:14 crc kubenswrapper[4926]: I1125 09:54:14.016557 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:54:14 crc kubenswrapper[4926]: I1125 09:54:14.017194 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:54:44 crc kubenswrapper[4926]: I1125 09:54:44.015686 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:54:44 crc kubenswrapper[4926]: I1125 09:54:44.017144 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.013747 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.014306 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.018962 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.019682 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2e7ab6176895234db45534a987817a5aad23e28ab3b2e4096d271630a3d44b38"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.019769 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://2e7ab6176895234db45534a987817a5aad23e28ab3b2e4096d271630a3d44b38" gracePeriod=600 Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.352540 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="2e7ab6176895234db45534a987817a5aad23e28ab3b2e4096d271630a3d44b38" exitCode=0 Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.352638 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"2e7ab6176895234db45534a987817a5aad23e28ab3b2e4096d271630a3d44b38"} Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.352882 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"cbade86616f5a9cb6f1541e747cc180bfbb8672001b5ece28519dd37aebd3ef3"} Nov 25 09:55:14 crc kubenswrapper[4926]: I1125 09:55:14.352909 4926 scope.go:117] "RemoveContainer" containerID="e38b5fc111cb4ebe271a7800e54ac04cd3878767af21589f11433f4e2f8cd78d" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.180832 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zvms9"] Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.182191 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.194270 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zvms9"] Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.350482 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-bound-sa-token\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.351060 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzct7\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-kube-api-access-bzct7\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.351118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.351863 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-registry-tls\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.351932 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ceaf0c2c-518f-4175-913f-07389fe657d3-trusted-ca\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.352112 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ceaf0c2c-518f-4175-913f-07389fe657d3-registry-certificates\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.352266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ceaf0c2c-518f-4175-913f-07389fe657d3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.352366 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ceaf0c2c-518f-4175-913f-07389fe657d3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.375020 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453095 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-registry-tls\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453166 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ceaf0c2c-518f-4175-913f-07389fe657d3-registry-certificates\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453187 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ceaf0c2c-518f-4175-913f-07389fe657d3-trusted-ca\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453221 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ceaf0c2c-518f-4175-913f-07389fe657d3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453252 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ceaf0c2c-518f-4175-913f-07389fe657d3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453295 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-bound-sa-token\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.453314 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzct7\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-kube-api-access-bzct7\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.454204 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/ceaf0c2c-518f-4175-913f-07389fe657d3-ca-trust-extracted\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.455454 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/ceaf0c2c-518f-4175-913f-07389fe657d3-registry-certificates\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.455606 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ceaf0c2c-518f-4175-913f-07389fe657d3-trusted-ca\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.461419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/ceaf0c2c-518f-4175-913f-07389fe657d3-installation-pull-secrets\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.463418 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-registry-tls\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.472697 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-bound-sa-token\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.474438 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzct7\" (UniqueName: \"kubernetes.io/projected/ceaf0c2c-518f-4175-913f-07389fe657d3-kube-api-access-bzct7\") pod \"image-registry-66df7c8f76-zvms9\" (UID: \"ceaf0c2c-518f-4175-913f-07389fe657d3\") " pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.499832 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:49 crc kubenswrapper[4926]: I1125 09:56:49.995829 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-zvms9"] Nov 25 09:56:51 crc kubenswrapper[4926]: I1125 09:56:51.008133 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" event={"ID":"ceaf0c2c-518f-4175-913f-07389fe657d3","Type":"ContainerStarted","Data":"7be277ba1350444a2eeb57a8e3c2dca324d58f48b4be702fa992691707232105"} Nov 25 09:56:51 crc kubenswrapper[4926]: I1125 09:56:51.008487 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:56:51 crc kubenswrapper[4926]: I1125 09:56:51.008502 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" event={"ID":"ceaf0c2c-518f-4175-913f-07389fe657d3","Type":"ContainerStarted","Data":"4d983b004315d791663867d84fea32078a2ab9fb96c5d262060a52dd3bdf04f8"} Nov 25 09:56:51 crc kubenswrapper[4926]: I1125 09:56:51.039054 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" podStartSLOduration=2.039027944 podStartE2EDuration="2.039027944s" podCreationTimestamp="2025-11-25 09:56:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:56:51.032595567 +0000 UTC m=+507.716551984" watchObservedRunningTime="2025-11-25 09:56:51.039027944 +0000 UTC m=+507.722984391" Nov 25 09:57:09 crc kubenswrapper[4926]: I1125 09:57:09.508098 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" Nov 25 09:57:09 crc kubenswrapper[4926]: I1125 09:57:09.583795 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zltj"] Nov 25 09:57:14 crc kubenswrapper[4926]: I1125 09:57:14.014289 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:57:14 crc kubenswrapper[4926]: I1125 09:57:14.014646 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:57:34 crc kubenswrapper[4926]: I1125 09:57:34.633551 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" podUID="36216347-d1f5-4db3-8bb8-6654caba4839" containerName="registry" containerID="cri-o://7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb" gracePeriod=30 Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.043837 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153478 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsc7s\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-kube-api-access-rsc7s\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153540 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-bound-sa-token\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153714 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153761 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-trusted-ca\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153800 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/36216347-d1f5-4db3-8bb8-6654caba4839-installation-pull-secrets\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153835 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-registry-certificates\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.153918 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-registry-tls\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.154584 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.154600 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.154689 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/36216347-d1f5-4db3-8bb8-6654caba4839-ca-trust-extracted\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.156478 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.156526 4926 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/36216347-d1f5-4db3-8bb8-6654caba4839-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.159807 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36216347-d1f5-4db3-8bb8-6654caba4839-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.159810 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.160660 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.161623 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-kube-api-access-rsc7s" (OuterVolumeSpecName: "kube-api-access-rsc7s") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "kube-api-access-rsc7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: E1125 09:57:35.162025 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:36216347-d1f5-4db3-8bb8-6654caba4839 nodeName:}" failed. No retries permitted until 2025-11-25 09:57:35.661975999 +0000 UTC m=+552.345932436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "registry-storage" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839") : kubernetes.io/csi: Unmounter.TearDownAt failed: rpc error: code = Unknown desc = check target path: could not get consistent content of /proc/mounts after 3 attempts Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.170202 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/36216347-d1f5-4db3-8bb8-6654caba4839-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.257846 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsc7s\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-kube-api-access-rsc7s\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.257913 4926 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.257922 4926 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/36216347-d1f5-4db3-8bb8-6654caba4839-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.257931 4926 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/36216347-d1f5-4db3-8bb8-6654caba4839-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.257940 4926 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/36216347-d1f5-4db3-8bb8-6654caba4839-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.330384 4926 generic.go:334] "Generic (PLEG): container finished" podID="36216347-d1f5-4db3-8bb8-6654caba4839" containerID="7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb" exitCode=0 Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.330433 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" event={"ID":"36216347-d1f5-4db3-8bb8-6654caba4839","Type":"ContainerDied","Data":"7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb"} Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.330489 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" event={"ID":"36216347-d1f5-4db3-8bb8-6654caba4839","Type":"ContainerDied","Data":"31b51098160244cad5ffd4c299c03bc7dd70de57e15783e53acade310ff42fe8"} Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.330515 4926 scope.go:117] "RemoveContainer" containerID="7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.330448 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-8zltj" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.374138 4926 scope.go:117] "RemoveContainer" containerID="7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb" Nov 25 09:57:35 crc kubenswrapper[4926]: E1125 09:57:35.374629 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb\": container with ID starting with 7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb not found: ID does not exist" containerID="7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.374666 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb"} err="failed to get container status \"7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb\": rpc error: code = NotFound desc = could not find container \"7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb\": container with ID starting with 7a3d8451bc876a9e22990836b50afbad5d774dcda93519eabc7ec90779f9f9fb not found: ID does not exist" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.664367 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"36216347-d1f5-4db3-8bb8-6654caba4839\" (UID: \"36216347-d1f5-4db3-8bb8-6654caba4839\") " Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.679236 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "36216347-d1f5-4db3-8bb8-6654caba4839" (UID: "36216347-d1f5-4db3-8bb8-6654caba4839"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.778612 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zltj"] Nov 25 09:57:35 crc kubenswrapper[4926]: I1125 09:57:35.789526 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-8zltj"] Nov 25 09:57:36 crc kubenswrapper[4926]: I1125 09:57:36.018894 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36216347-d1f5-4db3-8bb8-6654caba4839" path="/var/lib/kubelet/pods/36216347-d1f5-4db3-8bb8-6654caba4839/volumes" Nov 25 09:57:44 crc kubenswrapper[4926]: I1125 09:57:44.014465 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:57:44 crc kubenswrapper[4926]: I1125 09:57:44.015252 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.014986 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.015561 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.019131 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.019666 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cbade86616f5a9cb6f1541e747cc180bfbb8672001b5ece28519dd37aebd3ef3"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.019730 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://cbade86616f5a9cb6f1541e747cc180bfbb8672001b5ece28519dd37aebd3ef3" gracePeriod=600 Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.588037 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="cbade86616f5a9cb6f1541e747cc180bfbb8672001b5ece28519dd37aebd3ef3" exitCode=0 Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.588110 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"cbade86616f5a9cb6f1541e747cc180bfbb8672001b5ece28519dd37aebd3ef3"} Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.588517 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"d32c8b7deb1af6488ea2af61857b46c8de27ff7dc9d3d581bdb4b8b8f28e8724"} Nov 25 09:58:14 crc kubenswrapper[4926]: I1125 09:58:14.588559 4926 scope.go:117] "RemoveContainer" containerID="2e7ab6176895234db45534a987817a5aad23e28ab3b2e4096d271630a3d44b38" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.836858 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hpkft"] Nov 25 09:59:12 crc kubenswrapper[4926]: E1125 09:59:12.837798 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36216347-d1f5-4db3-8bb8-6654caba4839" containerName="registry" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.837816 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="36216347-d1f5-4db3-8bb8-6654caba4839" containerName="registry" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.837954 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="36216347-d1f5-4db3-8bb8-6654caba4839" containerName="registry" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.838460 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.841261 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-5ntjb" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.841577 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.841605 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.849826 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hpkft"] Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.867690 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8pj2r"] Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.868974 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.871744 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mx46t" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.873112 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-sxxx6"] Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.873834 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.878642 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vmcx9" Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.909952 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8pj2r"] Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.916207 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-sxxx6"] Nov 25 09:59:12 crc kubenswrapper[4926]: I1125 09:59:12.930449 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm7z7\" (UniqueName: \"kubernetes.io/projected/96a34dc4-fd0b-43c2-bed3-741482089f7b-kube-api-access-xm7z7\") pod \"cert-manager-cainjector-7f985d654d-hpkft\" (UID: \"96a34dc4-fd0b-43c2-bed3-741482089f7b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.032155 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n44lf\" (UniqueName: \"kubernetes.io/projected/1fddcfdd-0fb9-4153-a227-b28e6faccb40-kube-api-access-n44lf\") pod \"cert-manager-5b446d88c5-sxxx6\" (UID: \"1fddcfdd-0fb9-4153-a227-b28e6faccb40\") " pod="cert-manager/cert-manager-5b446d88c5-sxxx6" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.032266 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm7z7\" (UniqueName: \"kubernetes.io/projected/96a34dc4-fd0b-43c2-bed3-741482089f7b-kube-api-access-xm7z7\") pod \"cert-manager-cainjector-7f985d654d-hpkft\" (UID: \"96a34dc4-fd0b-43c2-bed3-741482089f7b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.032302 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbcw9\" (UniqueName: \"kubernetes.io/projected/6b70e250-e5c4-4754-b5d1-a4fe76f60b00-kube-api-access-sbcw9\") pod \"cert-manager-webhook-5655c58dd6-8pj2r\" (UID: \"6b70e250-e5c4-4754-b5d1-a4fe76f60b00\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.050161 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm7z7\" (UniqueName: \"kubernetes.io/projected/96a34dc4-fd0b-43c2-bed3-741482089f7b-kube-api-access-xm7z7\") pod \"cert-manager-cainjector-7f985d654d-hpkft\" (UID: \"96a34dc4-fd0b-43c2-bed3-741482089f7b\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.133427 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbcw9\" (UniqueName: \"kubernetes.io/projected/6b70e250-e5c4-4754-b5d1-a4fe76f60b00-kube-api-access-sbcw9\") pod \"cert-manager-webhook-5655c58dd6-8pj2r\" (UID: \"6b70e250-e5c4-4754-b5d1-a4fe76f60b00\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.133516 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n44lf\" (UniqueName: \"kubernetes.io/projected/1fddcfdd-0fb9-4153-a227-b28e6faccb40-kube-api-access-n44lf\") pod \"cert-manager-5b446d88c5-sxxx6\" (UID: \"1fddcfdd-0fb9-4153-a227-b28e6faccb40\") " pod="cert-manager/cert-manager-5b446d88c5-sxxx6" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.155312 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.155721 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n44lf\" (UniqueName: \"kubernetes.io/projected/1fddcfdd-0fb9-4153-a227-b28e6faccb40-kube-api-access-n44lf\") pod \"cert-manager-5b446d88c5-sxxx6\" (UID: \"1fddcfdd-0fb9-4153-a227-b28e6faccb40\") " pod="cert-manager/cert-manager-5b446d88c5-sxxx6" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.161100 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbcw9\" (UniqueName: \"kubernetes.io/projected/6b70e250-e5c4-4754-b5d1-a4fe76f60b00-kube-api-access-sbcw9\") pod \"cert-manager-webhook-5655c58dd6-8pj2r\" (UID: \"6b70e250-e5c4-4754-b5d1-a4fe76f60b00\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.200881 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.214608 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.441144 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-hpkft"] Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.456633 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.521038 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-8pj2r"] Nov 25 09:59:13 crc kubenswrapper[4926]: W1125 09:59:13.529508 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b70e250_e5c4_4754_b5d1_a4fe76f60b00.slice/crio-914af9c67a89b384a2a6d0a6c0be193764571d6feff75413cfa269fbcfa2629a WatchSource:0}: Error finding container 914af9c67a89b384a2a6d0a6c0be193764571d6feff75413cfa269fbcfa2629a: Status 404 returned error can't find the container with id 914af9c67a89b384a2a6d0a6c0be193764571d6feff75413cfa269fbcfa2629a Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.550616 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-sxxx6"] Nov 25 09:59:13 crc kubenswrapper[4926]: W1125 09:59:13.554651 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fddcfdd_0fb9_4153_a227_b28e6faccb40.slice/crio-812eb1bde8e989aaf944e41b2a9f6ad1c9177c1a97eeebb66ea0b41ea7ce737c WatchSource:0}: Error finding container 812eb1bde8e989aaf944e41b2a9f6ad1c9177c1a97eeebb66ea0b41ea7ce737c: Status 404 returned error can't find the container with id 812eb1bde8e989aaf944e41b2a9f6ad1c9177c1a97eeebb66ea0b41ea7ce737c Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.945827 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" event={"ID":"1fddcfdd-0fb9-4153-a227-b28e6faccb40","Type":"ContainerStarted","Data":"812eb1bde8e989aaf944e41b2a9f6ad1c9177c1a97eeebb66ea0b41ea7ce737c"} Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.947281 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" event={"ID":"6b70e250-e5c4-4754-b5d1-a4fe76f60b00","Type":"ContainerStarted","Data":"914af9c67a89b384a2a6d0a6c0be193764571d6feff75413cfa269fbcfa2629a"} Nov 25 09:59:13 crc kubenswrapper[4926]: I1125 09:59:13.948422 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" event={"ID":"96a34dc4-fd0b-43c2-bed3-741482089f7b","Type":"ContainerStarted","Data":"75e8a7a85bdbc01178f45598833d26f2cbb822c3028b4348af35ea638e0b02be"} Nov 25 09:59:17 crc kubenswrapper[4926]: I1125 09:59:17.983704 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" event={"ID":"6b70e250-e5c4-4754-b5d1-a4fe76f60b00","Type":"ContainerStarted","Data":"052531888df598e4624f184e517ff16868181fdd78f323da4b4c899233be0294"} Nov 25 09:59:17 crc kubenswrapper[4926]: I1125 09:59:17.984276 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:17 crc kubenswrapper[4926]: I1125 09:59:17.985596 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" event={"ID":"96a34dc4-fd0b-43c2-bed3-741482089f7b","Type":"ContainerStarted","Data":"401863c86ff6640ad7310a29bb44937ce36860976b9e7f5e2f82be773fc50970"} Nov 25 09:59:17 crc kubenswrapper[4926]: I1125 09:59:17.987457 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" event={"ID":"1fddcfdd-0fb9-4153-a227-b28e6faccb40","Type":"ContainerStarted","Data":"cebb5b05bd3519f415c184b376d12a6d38734df3ea33fc71629be18ee5fb573e"} Nov 25 09:59:18 crc kubenswrapper[4926]: I1125 09:59:18.005182 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" podStartSLOduration=2.485505911 podStartE2EDuration="6.005157964s" podCreationTimestamp="2025-11-25 09:59:12 +0000 UTC" firstStartedPulling="2025-11-25 09:59:13.531983086 +0000 UTC m=+650.215939513" lastFinishedPulling="2025-11-25 09:59:17.051635089 +0000 UTC m=+653.735591566" observedRunningTime="2025-11-25 09:59:18.001461223 +0000 UTC m=+654.685417680" watchObservedRunningTime="2025-11-25 09:59:18.005157964 +0000 UTC m=+654.689114391" Nov 25 09:59:18 crc kubenswrapper[4926]: I1125 09:59:18.026747 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" podStartSLOduration=2.54870708 podStartE2EDuration="6.026718893s" podCreationTimestamp="2025-11-25 09:59:12 +0000 UTC" firstStartedPulling="2025-11-25 09:59:13.55872297 +0000 UTC m=+650.242679387" lastFinishedPulling="2025-11-25 09:59:17.036734753 +0000 UTC m=+653.720691200" observedRunningTime="2025-11-25 09:59:18.025438592 +0000 UTC m=+654.709395049" watchObservedRunningTime="2025-11-25 09:59:18.026718893 +0000 UTC m=+654.710675330" Nov 25 09:59:18 crc kubenswrapper[4926]: I1125 09:59:18.048818 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" podStartSLOduration=2.4676689720000002 podStartE2EDuration="6.048793005s" podCreationTimestamp="2025-11-25 09:59:12 +0000 UTC" firstStartedPulling="2025-11-25 09:59:13.456291037 +0000 UTC m=+650.140247464" lastFinishedPulling="2025-11-25 09:59:17.03741506 +0000 UTC m=+653.721371497" observedRunningTime="2025-11-25 09:59:18.043392233 +0000 UTC m=+654.727348680" watchObservedRunningTime="2025-11-25 09:59:18.048793005 +0000 UTC m=+654.732749442" Nov 25 09:59:23 crc kubenswrapper[4926]: I1125 09:59:23.203701 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-8pj2r" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.476063 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4pcsz"] Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477122 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-controller" containerID="cri-o://c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477594 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="sbdb" containerID="cri-o://e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477643 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="nbdb" containerID="cri-o://c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477685 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="northd" containerID="cri-o://7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477723 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477762 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-node" containerID="cri-o://aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.477800 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-acl-logging" containerID="cri-o://6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.513752 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" containerID="cri-o://f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" gracePeriod=30 Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.810523 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/3.log" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.813284 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovn-acl-logging/0.log" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.813968 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovn-controller/0.log" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.814579 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.876999 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-5kwrg"] Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877250 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="northd" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877267 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="northd" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877279 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877285 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877296 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877303 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877318 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="sbdb" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877324 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="sbdb" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877333 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kubecfg-setup" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877339 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kubecfg-setup" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877348 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-node" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877353 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-node" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877360 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877365 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877374 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877380 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877387 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877393 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877401 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="nbdb" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877407 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="nbdb" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877416 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-acl-logging" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877422 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-acl-logging" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877432 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877437 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877527 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877537 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877548 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-node" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877556 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="nbdb" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877565 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovn-acl-logging" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877574 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877583 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="kube-rbac-proxy-ovn-metrics" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877591 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877600 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877606 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="sbdb" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877617 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="northd" Nov 25 09:59:45 crc kubenswrapper[4926]: E1125 09:59:45.877706 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877713 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.877811 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerName="ovnkube-controller" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.879604 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916000 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-ovn-kubernetes\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916051 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-systemd-units\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916074 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-kubelet\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916087 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916112 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-netns\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916162 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916201 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916201 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-node-log\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916222 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-node-log" (OuterVolumeSpecName: "node-log") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916226 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-var-lib-openvswitch\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916243 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-openvswitch\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916247 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916261 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-bin\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916269 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916286 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-log-socket\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916304 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916308 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-config\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916339 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-systemd\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916360 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-etc-openvswitch\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916401 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-env-overrides\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916438 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-netd\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916460 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-ovn\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916481 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtqrw\" (UniqueName: \"kubernetes.io/projected/9a4d98c7-0970-4ab3-86a6-40c61350f745-kube-api-access-xtqrw\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916523 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-script-lib\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916547 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-var-lib-cni-networks-ovn-kubernetes\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916609 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-slash\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916634 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovn-node-metrics-cert\") pod \"9a4d98c7-0970-4ab3-86a6-40c61350f745\" (UID: \"9a4d98c7-0970-4ab3-86a6-40c61350f745\") " Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916809 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-etc-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovnkube-config\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916880 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-log-socket\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916909 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-systemd-units\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916929 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovn-node-metrics-cert\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917042 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-ovn\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917059 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-env-overrides\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916810 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917206 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-node-log\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917243 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-cni-bin\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917271 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-kubelet\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917296 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917342 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917362 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq587\" (UniqueName: \"kubernetes.io/projected/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-kube-api-access-zq587\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-slash" (OuterVolumeSpecName: "host-slash") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917394 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916831 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.916844 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-log-socket" (OuterVolumeSpecName: "log-socket") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917423 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917447 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-systemd\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917460 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-slash\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-cni-netd\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917583 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovnkube-script-lib\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917701 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917729 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-run-netns\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917746 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917796 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-var-lib-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917929 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.917981 4926 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918000 4926 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-slash\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918012 4926 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918025 4926 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918035 4926 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918046 4926 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918056 4926 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-node-log\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918066 4926 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918074 4926 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918084 4926 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918093 4926 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-log-socket\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918102 4926 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918114 4926 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918123 4926 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/9a4d98c7-0970-4ab3-86a6-40c61350f745-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918132 4926 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.918141 4926 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.923148 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a4d98c7-0970-4ab3-86a6-40c61350f745-kube-api-access-xtqrw" (OuterVolumeSpecName: "kube-api-access-xtqrw") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "kube-api-access-xtqrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.923652 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 09:59:45 crc kubenswrapper[4926]: I1125 09:59:45.930539 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "9a4d98c7-0970-4ab3-86a6-40c61350f745" (UID: "9a4d98c7-0970-4ab3-86a6-40c61350f745"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019226 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-run-netns\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019309 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-run-netns\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019359 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019420 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-var-lib-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019440 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-run-ovn-kubernetes\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019484 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-etc-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019507 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovnkube-config\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019528 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-log-socket\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019539 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-etc-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-systemd-units\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019560 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-systemd-units\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019487 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-var-lib-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovn-node-metrics-cert\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019723 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-ovn\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-env-overrides\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019835 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-node-log\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019882 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-cni-bin\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019923 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-kubelet\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.019988 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020076 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq587\" (UniqueName: \"kubernetes.io/projected/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-kube-api-access-zq587\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020154 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-systemd\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020246 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-slash\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020316 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-node-log\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020341 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-cni-netd\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-cni-bin\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020430 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovnkube-script-lib\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020529 4926 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/9a4d98c7-0970-4ab3-86a6-40c61350f745-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020552 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtqrw\" (UniqueName: \"kubernetes.io/projected/9a4d98c7-0970-4ab3-86a6-40c61350f745-kube-api-access-xtqrw\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020574 4926 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/9a4d98c7-0970-4ab3-86a6-40c61350f745-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020643 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-kubelet\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020717 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-slash\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020755 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-cni-netd\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020295 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovnkube-config\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020806 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-ovn\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020817 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-systemd\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.020826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-run-openvswitch\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.021030 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-log-socket\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.021266 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-env-overrides\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.021563 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovnkube-script-lib\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.025888 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-ovn-node-metrics-cert\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.045406 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq587\" (UniqueName: \"kubernetes.io/projected/a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3-kube-api-access-zq587\") pod \"ovnkube-node-5kwrg\" (UID: \"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3\") " pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.173619 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/2.log" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.174121 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/1.log" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.174183 4926 generic.go:334] "Generic (PLEG): container finished" podID="cab27fcb-cde4-4879-8fad-79951474d7b3" containerID="e52d67a89b7079554fcf1c05dd1fccff682bba2a9a44c7601776d730e21d6fe6" exitCode=2 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.174291 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerDied","Data":"e52d67a89b7079554fcf1c05dd1fccff682bba2a9a44c7601776d730e21d6fe6"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.174428 4926 scope.go:117] "RemoveContainer" containerID="1fc226654b0ae6fbad5920f756bf7fe7b5d3fa09b64e1d684709f5044b1abf54" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.175043 4926 scope.go:117] "RemoveContainer" containerID="e52d67a89b7079554fcf1c05dd1fccff682bba2a9a44c7601776d730e21d6fe6" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.175321 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-l78n4_openshift-multus(cab27fcb-cde4-4879-8fad-79951474d7b3)\"" pod="openshift-multus/multus-l78n4" podUID="cab27fcb-cde4-4879-8fad-79951474d7b3" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.178643 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovnkube-controller/3.log" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.181475 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovn-acl-logging/0.log" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182005 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-4pcsz_9a4d98c7-0970-4ab3-86a6-40c61350f745/ovn-controller/0.log" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182449 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" exitCode=0 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182511 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182569 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182585 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182530 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" exitCode=0 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182621 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" exitCode=0 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182637 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" exitCode=0 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182646 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" exitCode=0 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182653 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" exitCode=0 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182661 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" exitCode=143 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182670 4926 generic.go:334] "Generic (PLEG): container finished" podID="9a4d98c7-0970-4ab3-86a6-40c61350f745" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" exitCode=143 Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182688 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182701 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182716 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182749 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182766 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182773 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182781 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182788 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182794 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182802 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182809 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182816 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182823 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182833 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182845 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182853 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182861 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182868 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182875 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182881 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182890 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182897 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182903 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182912 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182923 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182971 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182983 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182992 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.182999 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183007 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183015 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183024 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183032 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183039 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183046 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183067 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-4pcsz" event={"ID":"9a4d98c7-0970-4ab3-86a6-40c61350f745","Type":"ContainerDied","Data":"0e11357508f6dabaac0c0c3c16c6522bb4481cb60be7fdd562912280fd2bad02"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183081 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183088 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183095 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183102 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183108 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183114 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183121 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183127 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183134 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.183140 4926 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.193889 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.204233 4926 scope.go:117] "RemoveContainer" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.227570 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4pcsz"] Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.230680 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-4pcsz"] Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.237314 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.259049 4926 scope.go:117] "RemoveContainer" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.277849 4926 scope.go:117] "RemoveContainer" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.297083 4926 scope.go:117] "RemoveContainer" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.319320 4926 scope.go:117] "RemoveContainer" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.339994 4926 scope.go:117] "RemoveContainer" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.355538 4926 scope.go:117] "RemoveContainer" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.370287 4926 scope.go:117] "RemoveContainer" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.385468 4926 scope.go:117] "RemoveContainer" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.452258 4926 scope.go:117] "RemoveContainer" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.452768 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": container with ID starting with f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f not found: ID does not exist" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.452819 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} err="failed to get container status \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": rpc error: code = NotFound desc = could not find container \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": container with ID starting with f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.452859 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.453330 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": container with ID starting with 71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b not found: ID does not exist" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.453378 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} err="failed to get container status \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": rpc error: code = NotFound desc = could not find container \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": container with ID starting with 71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.453416 4926 scope.go:117] "RemoveContainer" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.453663 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": container with ID starting with e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda not found: ID does not exist" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.453693 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} err="failed to get container status \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": rpc error: code = NotFound desc = could not find container \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": container with ID starting with e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.453712 4926 scope.go:117] "RemoveContainer" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.454264 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": container with ID starting with c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c not found: ID does not exist" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.454291 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} err="failed to get container status \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": rpc error: code = NotFound desc = could not find container \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": container with ID starting with c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.454310 4926 scope.go:117] "RemoveContainer" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.454823 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": container with ID starting with 7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2 not found: ID does not exist" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.454863 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} err="failed to get container status \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": rpc error: code = NotFound desc = could not find container \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": container with ID starting with 7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.454883 4926 scope.go:117] "RemoveContainer" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.455164 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": container with ID starting with ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c not found: ID does not exist" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.455184 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} err="failed to get container status \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": rpc error: code = NotFound desc = could not find container \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": container with ID starting with ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.455200 4926 scope.go:117] "RemoveContainer" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.455631 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": container with ID starting with aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d not found: ID does not exist" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.455681 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} err="failed to get container status \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": rpc error: code = NotFound desc = could not find container \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": container with ID starting with aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.455723 4926 scope.go:117] "RemoveContainer" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.456094 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": container with ID starting with 6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9 not found: ID does not exist" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.456141 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} err="failed to get container status \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": rpc error: code = NotFound desc = could not find container \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": container with ID starting with 6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.456160 4926 scope.go:117] "RemoveContainer" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.456450 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": container with ID starting with c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e not found: ID does not exist" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.456474 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} err="failed to get container status \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": rpc error: code = NotFound desc = could not find container \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": container with ID starting with c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.456489 4926 scope.go:117] "RemoveContainer" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" Nov 25 09:59:46 crc kubenswrapper[4926]: E1125 09:59:46.456727 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": container with ID starting with 2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0 not found: ID does not exist" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.456750 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} err="failed to get container status \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": rpc error: code = NotFound desc = could not find container \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": container with ID starting with 2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.456765 4926 scope.go:117] "RemoveContainer" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457030 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} err="failed to get container status \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": rpc error: code = NotFound desc = could not find container \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": container with ID starting with f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457057 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457271 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} err="failed to get container status \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": rpc error: code = NotFound desc = could not find container \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": container with ID starting with 71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457297 4926 scope.go:117] "RemoveContainer" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457620 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} err="failed to get container status \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": rpc error: code = NotFound desc = could not find container \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": container with ID starting with e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457658 4926 scope.go:117] "RemoveContainer" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.457994 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} err="failed to get container status \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": rpc error: code = NotFound desc = could not find container \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": container with ID starting with c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458030 4926 scope.go:117] "RemoveContainer" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458289 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} err="failed to get container status \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": rpc error: code = NotFound desc = could not find container \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": container with ID starting with 7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458310 4926 scope.go:117] "RemoveContainer" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458530 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} err="failed to get container status \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": rpc error: code = NotFound desc = could not find container \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": container with ID starting with ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458557 4926 scope.go:117] "RemoveContainer" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458823 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} err="failed to get container status \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": rpc error: code = NotFound desc = could not find container \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": container with ID starting with aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.458842 4926 scope.go:117] "RemoveContainer" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.460349 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} err="failed to get container status \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": rpc error: code = NotFound desc = could not find container \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": container with ID starting with 6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.460383 4926 scope.go:117] "RemoveContainer" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.460740 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} err="failed to get container status \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": rpc error: code = NotFound desc = could not find container \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": container with ID starting with c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.460760 4926 scope.go:117] "RemoveContainer" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.461330 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} err="failed to get container status \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": rpc error: code = NotFound desc = could not find container \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": container with ID starting with 2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.461352 4926 scope.go:117] "RemoveContainer" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.461624 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} err="failed to get container status \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": rpc error: code = NotFound desc = could not find container \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": container with ID starting with f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.461660 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.462100 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} err="failed to get container status \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": rpc error: code = NotFound desc = could not find container \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": container with ID starting with 71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.462134 4926 scope.go:117] "RemoveContainer" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.462451 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} err="failed to get container status \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": rpc error: code = NotFound desc = could not find container \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": container with ID starting with e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.462477 4926 scope.go:117] "RemoveContainer" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.462806 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} err="failed to get container status \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": rpc error: code = NotFound desc = could not find container \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": container with ID starting with c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.462840 4926 scope.go:117] "RemoveContainer" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.463196 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} err="failed to get container status \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": rpc error: code = NotFound desc = could not find container \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": container with ID starting with 7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.463224 4926 scope.go:117] "RemoveContainer" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.463659 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} err="failed to get container status \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": rpc error: code = NotFound desc = could not find container \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": container with ID starting with ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.463693 4926 scope.go:117] "RemoveContainer" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.464389 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} err="failed to get container status \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": rpc error: code = NotFound desc = could not find container \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": container with ID starting with aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.464460 4926 scope.go:117] "RemoveContainer" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.464928 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} err="failed to get container status \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": rpc error: code = NotFound desc = could not find container \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": container with ID starting with 6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.464986 4926 scope.go:117] "RemoveContainer" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.465333 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} err="failed to get container status \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": rpc error: code = NotFound desc = could not find container \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": container with ID starting with c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.465362 4926 scope.go:117] "RemoveContainer" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.465688 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} err="failed to get container status \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": rpc error: code = NotFound desc = could not find container \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": container with ID starting with 2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.465725 4926 scope.go:117] "RemoveContainer" containerID="f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.466129 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f"} err="failed to get container status \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": rpc error: code = NotFound desc = could not find container \"f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f\": container with ID starting with f4c5e112fca2abb4bee244c1fdb24b3a7e8e67a90a4bbb60c7bef315e6025f7f not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.466152 4926 scope.go:117] "RemoveContainer" containerID="71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.466406 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b"} err="failed to get container status \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": rpc error: code = NotFound desc = could not find container \"71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b\": container with ID starting with 71f2716f7b0150d2db8efa0620e6d503a554f13a7afd62a39f1ab7a33ab4ba8b not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.466437 4926 scope.go:117] "RemoveContainer" containerID="e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.466670 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda"} err="failed to get container status \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": rpc error: code = NotFound desc = could not find container \"e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda\": container with ID starting with e22add33aa01695a4cb1593a452b1bc2f3a76794c3e0c77b1e059b3255ccbdda not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.466705 4926 scope.go:117] "RemoveContainer" containerID="c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.467021 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c"} err="failed to get container status \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": rpc error: code = NotFound desc = could not find container \"c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c\": container with ID starting with c24977477a49e036cba1ce5ad5766ada0f7ba336ab9985d09328c62dbef1d55c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.467042 4926 scope.go:117] "RemoveContainer" containerID="7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.472234 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2"} err="failed to get container status \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": rpc error: code = NotFound desc = could not find container \"7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2\": container with ID starting with 7a2f72b4912e1db4379fcc148f952f44393cc90039120d0e8fbb5943ff177de2 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.472291 4926 scope.go:117] "RemoveContainer" containerID="ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.475910 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c"} err="failed to get container status \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": rpc error: code = NotFound desc = could not find container \"ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c\": container with ID starting with ed0fe63bde520650bb474b7c367f8abae8cfb38262d1c63421379bd231ad633c not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.476196 4926 scope.go:117] "RemoveContainer" containerID="aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.476979 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d"} err="failed to get container status \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": rpc error: code = NotFound desc = could not find container \"aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d\": container with ID starting with aea8b82a58d7fc516e7bf63be46c140f1f5f6a425e105cfa2207e7ac6cb5403d not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.477007 4926 scope.go:117] "RemoveContainer" containerID="6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.477489 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9"} err="failed to get container status \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": rpc error: code = NotFound desc = could not find container \"6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9\": container with ID starting with 6c2292377afb4ac3c02e5dcbbb974e204627a1ced3c8ca8840394cf4f5bd12e9 not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.477583 4926 scope.go:117] "RemoveContainer" containerID="c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.478091 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e"} err="failed to get container status \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": rpc error: code = NotFound desc = could not find container \"c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e\": container with ID starting with c1a241cd00c82069d81ff36dfa64884d780d245ee3c1cefa1a2dcd3b0c01fe2e not found: ID does not exist" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.478112 4926 scope.go:117] "RemoveContainer" containerID="2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0" Nov 25 09:59:46 crc kubenswrapper[4926]: I1125 09:59:46.478490 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0"} err="failed to get container status \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": rpc error: code = NotFound desc = could not find container \"2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0\": container with ID starting with 2e7449824f8307eed2357408e46f14354ee6a81a03d4893dcabc09d438757ce0 not found: ID does not exist" Nov 25 09:59:47 crc kubenswrapper[4926]: I1125 09:59:47.192651 4926 generic.go:334] "Generic (PLEG): container finished" podID="a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3" containerID="9a2b57deb040ccf72e2c5d665207c05dd6771fce2a7c29b0b05f3f346549ecef" exitCode=0 Nov 25 09:59:47 crc kubenswrapper[4926]: I1125 09:59:47.192808 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerDied","Data":"9a2b57deb040ccf72e2c5d665207c05dd6771fce2a7c29b0b05f3f346549ecef"} Nov 25 09:59:47 crc kubenswrapper[4926]: I1125 09:59:47.192924 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"0bb687e41b82a0401fa209d01d46d0b6830168dc18594ef15d57ca1f9da491eb"} Nov 25 09:59:47 crc kubenswrapper[4926]: I1125 09:59:47.197820 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/2.log" Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.024014 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a4d98c7-0970-4ab3-86a6-40c61350f745" path="/var/lib/kubelet/pods/9a4d98c7-0970-4ab3-86a6-40c61350f745/volumes" Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.209549 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"034215d6895b80ac6510e82bb501ef971055e5f5cf9d5e68a13da3728fef6b62"} Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.209970 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"171c85c38965e8d459ece143b11610dc187a13c71edc664d2664ebb2d54eb9a2"} Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.209984 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"ab7a63df2d0dc6e012d26063f37f004b6e9b8b211477d40e4d00471af5995f60"} Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.209997 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"ae4274e04272a4f3a9b2166e036ef363d95ec57a8b2cde0d7639655218fe51e4"} Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.210007 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"673161d9423b9a552e5403281edd0e1e5ec7bc43b6428c0ca3d5cc26971ca964"} Nov 25 09:59:48 crc kubenswrapper[4926]: I1125 09:59:48.210019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"e2c6252d60dc8a503ef6b5121f286a9fb1491d5301fd2dd014ab25d6659b4d01"} Nov 25 09:59:51 crc kubenswrapper[4926]: I1125 09:59:51.250807 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"b54e1324bb3e7062d06246cc9518fca2db0c52be357661ae40d28daf3724bd8c"} Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.269458 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" event={"ID":"a1f00ee7-d6ba-4d3e-a145-39c3e3f133c3","Type":"ContainerStarted","Data":"c7cb0447ead07f0c749c072eb7b7de64e9a73c31ca57be0a2864dd0fd3ea9cc0"} Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.270072 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.270090 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.270105 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.302572 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" podStartSLOduration=8.302542164 podStartE2EDuration="8.302542164s" podCreationTimestamp="2025-11-25 09:59:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 09:59:53.299335835 +0000 UTC m=+689.983292302" watchObservedRunningTime="2025-11-25 09:59:53.302542164 +0000 UTC m=+689.986498591" Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.310845 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 09:59:53 crc kubenswrapper[4926]: I1125 09:59:53.318326 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.012706 4926 scope.go:117] "RemoveContainer" containerID="e52d67a89b7079554fcf1c05dd1fccff682bba2a9a44c7601776d730e21d6fe6" Nov 25 10:00:00 crc kubenswrapper[4926]: E1125 10:00:00.013610 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-l78n4_openshift-multus(cab27fcb-cde4-4879-8fad-79951474d7b3)\"" pod="openshift-multus/multus-l78n4" podUID="cab27fcb-cde4-4879-8fad-79951474d7b3" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.158766 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm"] Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.159869 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.162626 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.165978 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.168038 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm"] Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.241220 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d079f81-f2e0-4b4b-8102-952af5599033-secret-volume\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.241519 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx49b\" (UniqueName: \"kubernetes.io/projected/7d079f81-f2e0-4b4b-8102-952af5599033-kube-api-access-bx49b\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.241538 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d079f81-f2e0-4b4b-8102-952af5599033-config-volume\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.342465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d079f81-f2e0-4b4b-8102-952af5599033-secret-volume\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.342534 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx49b\" (UniqueName: \"kubernetes.io/projected/7d079f81-f2e0-4b4b-8102-952af5599033-kube-api-access-bx49b\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.342559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d079f81-f2e0-4b4b-8102-952af5599033-config-volume\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.344412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d079f81-f2e0-4b4b-8102-952af5599033-config-volume\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.353301 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d079f81-f2e0-4b4b-8102-952af5599033-secret-volume\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.361587 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx49b\" (UniqueName: \"kubernetes.io/projected/7d079f81-f2e0-4b4b-8102-952af5599033-kube-api-access-bx49b\") pod \"collect-profiles-29401080-2gtnm\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: I1125 10:00:00.480255 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: E1125 10:00:00.506877 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(2b0c8f8580f5c1eba4610df9ae3ec244a8db4cfa5904af3c1de448608d847cc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:00:00 crc kubenswrapper[4926]: E1125 10:00:00.506999 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(2b0c8f8580f5c1eba4610df9ae3ec244a8db4cfa5904af3c1de448608d847cc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: E1125 10:00:00.507028 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(2b0c8f8580f5c1eba4610df9ae3ec244a8db4cfa5904af3c1de448608d847cc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:00 crc kubenswrapper[4926]: E1125 10:00:00.507099 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager(7d079f81-f2e0-4b4b-8102-952af5599033)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager(7d079f81-f2e0-4b4b-8102-952af5599033)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(2b0c8f8580f5c1eba4610df9ae3ec244a8db4cfa5904af3c1de448608d847cc2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" podUID="7d079f81-f2e0-4b4b-8102-952af5599033" Nov 25 10:00:01 crc kubenswrapper[4926]: I1125 10:00:01.316623 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:01 crc kubenswrapper[4926]: I1125 10:00:01.317146 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:01 crc kubenswrapper[4926]: E1125 10:00:01.345083 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(f00034a05f52edfb12b51a1e4d0d11a67eb856b603b27754b9ddcae06fe9e4e6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:00:01 crc kubenswrapper[4926]: E1125 10:00:01.345164 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(f00034a05f52edfb12b51a1e4d0d11a67eb856b603b27754b9ddcae06fe9e4e6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:01 crc kubenswrapper[4926]: E1125 10:00:01.345191 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(f00034a05f52edfb12b51a1e4d0d11a67eb856b603b27754b9ddcae06fe9e4e6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:01 crc kubenswrapper[4926]: E1125 10:00:01.345250 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager(7d079f81-f2e0-4b4b-8102-952af5599033)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager(7d079f81-f2e0-4b4b-8102-952af5599033)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(f00034a05f52edfb12b51a1e4d0d11a67eb856b603b27754b9ddcae06fe9e4e6): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" podUID="7d079f81-f2e0-4b4b-8102-952af5599033" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.723795 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62"] Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.726546 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.731252 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.734820 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62"] Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.777374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.777424 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.777749 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7nk8\" (UniqueName: \"kubernetes.io/projected/2649872a-06ff-4653-98e3-2f51dd040ac7-kube-api-access-g7nk8\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.879468 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.879532 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.879559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7nk8\" (UniqueName: \"kubernetes.io/projected/2649872a-06ff-4653-98e3-2f51dd040ac7-kube-api-access-g7nk8\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.880259 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-bundle\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.880362 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-util\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:02 crc kubenswrapper[4926]: I1125 10:00:02.899900 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7nk8\" (UniqueName: \"kubernetes.io/projected/2649872a-06ff-4653-98e3-2f51dd040ac7-kube-api-access-g7nk8\") pod \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: I1125 10:00:03.056519 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.095641 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(13b8379e6dd268cff5cedacf8fc853ba455715135796092bd31199ef68ee13cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.095755 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(13b8379e6dd268cff5cedacf8fc853ba455715135796092bd31199ef68ee13cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.095789 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(13b8379e6dd268cff5cedacf8fc853ba455715135796092bd31199ef68ee13cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.095892 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace(2649872a-06ff-4653-98e3-2f51dd040ac7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace(2649872a-06ff-4653-98e3-2f51dd040ac7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(13b8379e6dd268cff5cedacf8fc853ba455715135796092bd31199ef68ee13cb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" Nov 25 10:00:03 crc kubenswrapper[4926]: I1125 10:00:03.328304 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: I1125 10:00:03.329126 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.360366 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(42b0c6017522c5670042be57305d7b16563288ea8d0d05da9286cb4d9f989bfe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.360455 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(42b0c6017522c5670042be57305d7b16563288ea8d0d05da9286cb4d9f989bfe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.360500 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(42b0c6017522c5670042be57305d7b16563288ea8d0d05da9286cb4d9f989bfe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:03 crc kubenswrapper[4926]: E1125 10:00:03.360571 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace(2649872a-06ff-4653-98e3-2f51dd040ac7)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace(2649872a-06ff-4653-98e3-2f51dd040ac7)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62_openshift-marketplace_2649872a-06ff-4653-98e3-2f51dd040ac7_0(42b0c6017522c5670042be57305d7b16563288ea8d0d05da9286cb4d9f989bfe): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" Nov 25 10:00:13 crc kubenswrapper[4926]: I1125 10:00:13.012590 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:13 crc kubenswrapper[4926]: I1125 10:00:13.013332 4926 scope.go:117] "RemoveContainer" containerID="e52d67a89b7079554fcf1c05dd1fccff682bba2a9a44c7601776d730e21d6fe6" Nov 25 10:00:13 crc kubenswrapper[4926]: I1125 10:00:13.013603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:13 crc kubenswrapper[4926]: E1125 10:00:13.051336 4926 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(858ad2539858d4a30ebdda7c81f1794ad962a3d1e058423d78a34646b06615ed): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 25 10:00:13 crc kubenswrapper[4926]: E1125 10:00:13.051716 4926 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(858ad2539858d4a30ebdda7c81f1794ad962a3d1e058423d78a34646b06615ed): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:13 crc kubenswrapper[4926]: E1125 10:00:13.051753 4926 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(858ad2539858d4a30ebdda7c81f1794ad962a3d1e058423d78a34646b06615ed): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:13 crc kubenswrapper[4926]: E1125 10:00:13.051835 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager(7d079f81-f2e0-4b4b-8102-952af5599033)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager(7d079f81-f2e0-4b4b-8102-952af5599033)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_collect-profiles-29401080-2gtnm_openshift-operator-lifecycle-manager_7d079f81-f2e0-4b4b-8102-952af5599033_0(858ad2539858d4a30ebdda7c81f1794ad962a3d1e058423d78a34646b06615ed): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" podUID="7d079f81-f2e0-4b4b-8102-952af5599033" Nov 25 10:00:13 crc kubenswrapper[4926]: I1125 10:00:13.399161 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-l78n4_cab27fcb-cde4-4879-8fad-79951474d7b3/kube-multus/2.log" Nov 25 10:00:13 crc kubenswrapper[4926]: I1125 10:00:13.399235 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-l78n4" event={"ID":"cab27fcb-cde4-4879-8fad-79951474d7b3","Type":"ContainerStarted","Data":"0ec4148530a8cae4cc0f59a45893fa97ae364220f08ca5081f46d8339e0b9a86"} Nov 25 10:00:14 crc kubenswrapper[4926]: I1125 10:00:14.014375 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:00:14 crc kubenswrapper[4926]: I1125 10:00:14.014751 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:00:16 crc kubenswrapper[4926]: I1125 10:00:16.231284 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-5kwrg" Nov 25 10:00:17 crc kubenswrapper[4926]: I1125 10:00:17.012131 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:17 crc kubenswrapper[4926]: I1125 10:00:17.012760 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:17 crc kubenswrapper[4926]: I1125 10:00:17.208121 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62"] Nov 25 10:00:17 crc kubenswrapper[4926]: I1125 10:00:17.425143 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" event={"ID":"2649872a-06ff-4653-98e3-2f51dd040ac7","Type":"ContainerStarted","Data":"1f0319df21740b4ee3b5edf5e8927140927acf870f78fb33709fb597acf8b7a3"} Nov 25 10:00:17 crc kubenswrapper[4926]: I1125 10:00:17.426341 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" event={"ID":"2649872a-06ff-4653-98e3-2f51dd040ac7","Type":"ContainerStarted","Data":"6e2214261ee0fb60b50a60f91a6f68d0d183399a1250bc49f673617653479588"} Nov 25 10:00:18 crc kubenswrapper[4926]: I1125 10:00:18.432745 4926 generic.go:334] "Generic (PLEG): container finished" podID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerID="1f0319df21740b4ee3b5edf5e8927140927acf870f78fb33709fb597acf8b7a3" exitCode=0 Nov 25 10:00:18 crc kubenswrapper[4926]: I1125 10:00:18.432803 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" event={"ID":"2649872a-06ff-4653-98e3-2f51dd040ac7","Type":"ContainerDied","Data":"1f0319df21740b4ee3b5edf5e8927140927acf870f78fb33709fb597acf8b7a3"} Nov 25 10:00:21 crc kubenswrapper[4926]: I1125 10:00:21.468487 4926 generic.go:334] "Generic (PLEG): container finished" podID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerID="6de1d8e7c618a9d9724d83bfddf9507af322f3b799ab84f825ee759dcf80a8d6" exitCode=0 Nov 25 10:00:21 crc kubenswrapper[4926]: I1125 10:00:21.468586 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" event={"ID":"2649872a-06ff-4653-98e3-2f51dd040ac7","Type":"ContainerDied","Data":"6de1d8e7c618a9d9724d83bfddf9507af322f3b799ab84f825ee759dcf80a8d6"} Nov 25 10:00:22 crc kubenswrapper[4926]: I1125 10:00:22.479079 4926 generic.go:334] "Generic (PLEG): container finished" podID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerID="0c2c8f0a41c3f3b8fb9934ca907cd81c33a51c4c6940792280ab54a92882c32f" exitCode=0 Nov 25 10:00:22 crc kubenswrapper[4926]: I1125 10:00:22.479218 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" event={"ID":"2649872a-06ff-4653-98e3-2f51dd040ac7","Type":"ContainerDied","Data":"0c2c8f0a41c3f3b8fb9934ca907cd81c33a51c4c6940792280ab54a92882c32f"} Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.745091 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.909521 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7nk8\" (UniqueName: \"kubernetes.io/projected/2649872a-06ff-4653-98e3-2f51dd040ac7-kube-api-access-g7nk8\") pod \"2649872a-06ff-4653-98e3-2f51dd040ac7\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.910070 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-bundle\") pod \"2649872a-06ff-4653-98e3-2f51dd040ac7\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.910334 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-util\") pod \"2649872a-06ff-4653-98e3-2f51dd040ac7\" (UID: \"2649872a-06ff-4653-98e3-2f51dd040ac7\") " Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.911199 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-bundle" (OuterVolumeSpecName: "bundle") pod "2649872a-06ff-4653-98e3-2f51dd040ac7" (UID: "2649872a-06ff-4653-98e3-2f51dd040ac7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.919558 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2649872a-06ff-4653-98e3-2f51dd040ac7-kube-api-access-g7nk8" (OuterVolumeSpecName: "kube-api-access-g7nk8") pod "2649872a-06ff-4653-98e3-2f51dd040ac7" (UID: "2649872a-06ff-4653-98e3-2f51dd040ac7"). InnerVolumeSpecName "kube-api-access-g7nk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:00:23 crc kubenswrapper[4926]: I1125 10:00:23.933970 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-util" (OuterVolumeSpecName: "util") pod "2649872a-06ff-4653-98e3-2f51dd040ac7" (UID: "2649872a-06ff-4653-98e3-2f51dd040ac7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:00:24 crc kubenswrapper[4926]: I1125 10:00:24.013277 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7nk8\" (UniqueName: \"kubernetes.io/projected/2649872a-06ff-4653-98e3-2f51dd040ac7-kube-api-access-g7nk8\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:24 crc kubenswrapper[4926]: I1125 10:00:24.013363 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:24 crc kubenswrapper[4926]: I1125 10:00:24.013390 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2649872a-06ff-4653-98e3-2f51dd040ac7-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:24 crc kubenswrapper[4926]: I1125 10:00:24.495991 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" event={"ID":"2649872a-06ff-4653-98e3-2f51dd040ac7","Type":"ContainerDied","Data":"6e2214261ee0fb60b50a60f91a6f68d0d183399a1250bc49f673617653479588"} Nov 25 10:00:24 crc kubenswrapper[4926]: I1125 10:00:24.498031 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e2214261ee0fb60b50a60f91a6f68d0d183399a1250bc49f673617653479588" Nov 25 10:00:24 crc kubenswrapper[4926]: I1125 10:00:24.496044 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5c796334424b8139919e908729ac8fe5c1f6e7b6bc33540f00b4f8772ejqp62" Nov 25 10:00:27 crc kubenswrapper[4926]: I1125 10:00:27.012349 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:27 crc kubenswrapper[4926]: I1125 10:00:27.013400 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:27 crc kubenswrapper[4926]: I1125 10:00:27.275332 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm"] Nov 25 10:00:27 crc kubenswrapper[4926]: I1125 10:00:27.519221 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" event={"ID":"7d079f81-f2e0-4b4b-8102-952af5599033","Type":"ContainerStarted","Data":"688e4c655ec731231722cfea0bac4b50649ebdd530a050c57943f4e6fc10d13e"} Nov 25 10:00:27 crc kubenswrapper[4926]: I1125 10:00:27.519316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" event={"ID":"7d079f81-f2e0-4b4b-8102-952af5599033","Type":"ContainerStarted","Data":"a09d152e7d430efc47c9ed0fee46a46eb9539bb5376aec18daf40428153ee667"} Nov 25 10:00:28 crc kubenswrapper[4926]: I1125 10:00:28.527368 4926 generic.go:334] "Generic (PLEG): container finished" podID="7d079f81-f2e0-4b4b-8102-952af5599033" containerID="688e4c655ec731231722cfea0bac4b50649ebdd530a050c57943f4e6fc10d13e" exitCode=0 Nov 25 10:00:28 crc kubenswrapper[4926]: I1125 10:00:28.527520 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" event={"ID":"7d079f81-f2e0-4b4b-8102-952af5599033","Type":"ContainerDied","Data":"688e4c655ec731231722cfea0bac4b50649ebdd530a050c57943f4e6fc10d13e"} Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.378971 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-qh5st"] Nov 25 10:00:29 crc kubenswrapper[4926]: E1125 10:00:29.379288 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="util" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.379313 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="util" Nov 25 10:00:29 crc kubenswrapper[4926]: E1125 10:00:29.379333 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="extract" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.379344 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="extract" Nov 25 10:00:29 crc kubenswrapper[4926]: E1125 10:00:29.379357 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="pull" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.379368 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="pull" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.379523 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2649872a-06ff-4653-98e3-2f51dd040ac7" containerName="extract" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.380209 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.383141 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bkprg" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.383843 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.384164 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.395637 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-qh5st"] Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.490421 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pp89k\" (UniqueName: \"kubernetes.io/projected/e766d24a-f1e5-4686-a6b3-539b906cdcfa-kube-api-access-pp89k\") pod \"nmstate-operator-557fdffb88-qh5st\" (UID: \"e766d24a-f1e5-4686-a6b3-539b906cdcfa\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.591884 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pp89k\" (UniqueName: \"kubernetes.io/projected/e766d24a-f1e5-4686-a6b3-539b906cdcfa-kube-api-access-pp89k\") pod \"nmstate-operator-557fdffb88-qh5st\" (UID: \"e766d24a-f1e5-4686-a6b3-539b906cdcfa\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.620979 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pp89k\" (UniqueName: \"kubernetes.io/projected/e766d24a-f1e5-4686-a6b3-539b906cdcfa-kube-api-access-pp89k\") pod \"nmstate-operator-557fdffb88-qh5st\" (UID: \"e766d24a-f1e5-4686-a6b3-539b906cdcfa\") " pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.743580 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.814704 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.903065 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bx49b\" (UniqueName: \"kubernetes.io/projected/7d079f81-f2e0-4b4b-8102-952af5599033-kube-api-access-bx49b\") pod \"7d079f81-f2e0-4b4b-8102-952af5599033\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.903124 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d079f81-f2e0-4b4b-8102-952af5599033-config-volume\") pod \"7d079f81-f2e0-4b4b-8102-952af5599033\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.903154 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d079f81-f2e0-4b4b-8102-952af5599033-secret-volume\") pod \"7d079f81-f2e0-4b4b-8102-952af5599033\" (UID: \"7d079f81-f2e0-4b4b-8102-952af5599033\") " Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.903979 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d079f81-f2e0-4b4b-8102-952af5599033-config-volume" (OuterVolumeSpecName: "config-volume") pod "7d079f81-f2e0-4b4b-8102-952af5599033" (UID: "7d079f81-f2e0-4b4b-8102-952af5599033"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.907223 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d079f81-f2e0-4b4b-8102-952af5599033-kube-api-access-bx49b" (OuterVolumeSpecName: "kube-api-access-bx49b") pod "7d079f81-f2e0-4b4b-8102-952af5599033" (UID: "7d079f81-f2e0-4b4b-8102-952af5599033"). InnerVolumeSpecName "kube-api-access-bx49b". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:00:29 crc kubenswrapper[4926]: I1125 10:00:29.907378 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d079f81-f2e0-4b4b-8102-952af5599033-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7d079f81-f2e0-4b4b-8102-952af5599033" (UID: "7d079f81-f2e0-4b4b-8102-952af5599033"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.006590 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bx49b\" (UniqueName: \"kubernetes.io/projected/7d079f81-f2e0-4b4b-8102-952af5599033-kube-api-access-bx49b\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.006633 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7d079f81-f2e0-4b4b-8102-952af5599033-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.006642 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7d079f81-f2e0-4b4b-8102-952af5599033-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.227180 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-557fdffb88-qh5st"] Nov 25 10:00:30 crc kubenswrapper[4926]: W1125 10:00:30.235245 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode766d24a_f1e5_4686_a6b3_539b906cdcfa.slice/crio-9977afdab3a9fd7b1e9a7dde3873d871eb8d0ad975458ed636036f975f2b72e2 WatchSource:0}: Error finding container 9977afdab3a9fd7b1e9a7dde3873d871eb8d0ad975458ed636036f975f2b72e2: Status 404 returned error can't find the container with id 9977afdab3a9fd7b1e9a7dde3873d871eb8d0ad975458ed636036f975f2b72e2 Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.541125 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" event={"ID":"e766d24a-f1e5-4686-a6b3-539b906cdcfa","Type":"ContainerStarted","Data":"9977afdab3a9fd7b1e9a7dde3873d871eb8d0ad975458ed636036f975f2b72e2"} Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.542857 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" event={"ID":"7d079f81-f2e0-4b4b-8102-952af5599033","Type":"ContainerDied","Data":"a09d152e7d430efc47c9ed0fee46a46eb9539bb5376aec18daf40428153ee667"} Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.542882 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a09d152e7d430efc47c9ed0fee46a46eb9539bb5376aec18daf40428153ee667" Nov 25 10:00:30 crc kubenswrapper[4926]: I1125 10:00:30.543004 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401080-2gtnm" Nov 25 10:00:32 crc kubenswrapper[4926]: I1125 10:00:32.553739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" event={"ID":"e766d24a-f1e5-4686-a6b3-539b906cdcfa","Type":"ContainerStarted","Data":"c49f25aae09bb7db36d29e44ddfe5f214814abca2d955a3a56c5a137d8a311e0"} Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.618256 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-557fdffb88-qh5st" podStartSLOduration=2.465660354 podStartE2EDuration="4.618224206s" podCreationTimestamp="2025-11-25 10:00:29 +0000 UTC" firstStartedPulling="2025-11-25 10:00:30.239431224 +0000 UTC m=+726.923387651" lastFinishedPulling="2025-11-25 10:00:32.391995076 +0000 UTC m=+729.075951503" observedRunningTime="2025-11-25 10:00:32.573907337 +0000 UTC m=+729.257863774" watchObservedRunningTime="2025-11-25 10:00:33.618224206 +0000 UTC m=+730.302180663" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.619563 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd"] Nov 25 10:00:33 crc kubenswrapper[4926]: E1125 10:00:33.619877 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d079f81-f2e0-4b4b-8102-952af5599033" containerName="collect-profiles" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.619908 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d079f81-f2e0-4b4b-8102-952af5599033" containerName="collect-profiles" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.620130 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d079f81-f2e0-4b4b-8102-952af5599033" containerName="collect-profiles" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.621029 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.622555 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-dklj7" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.631007 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.631956 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.634478 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.645635 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.658567 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-42kdg"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.659632 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.669748 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.756747 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhzvn\" (UniqueName: \"kubernetes.io/projected/11e4ce92-8234-4e31-a300-1896267b083c-kube-api-access-dhzvn\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.756809 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8d62\" (UniqueName: \"kubernetes.io/projected/a04b62fe-2ebd-4b20-b4b0-93038ead2692-kube-api-access-k8d62\") pod \"nmstate-webhook-6b89b748d8-cbvck\" (UID: \"a04b62fe-2ebd-4b20-b4b0-93038ead2692\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.756841 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzm89\" (UniqueName: \"kubernetes.io/projected/230299a9-48a6-405b-9569-33b7110107be-kube-api-access-pzm89\") pod \"nmstate-metrics-5dcf9c57c5-lbjdd\" (UID: \"230299a9-48a6-405b-9569-33b7110107be\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.756888 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-dbus-socket\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.757081 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a04b62fe-2ebd-4b20-b4b0-93038ead2692-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-cbvck\" (UID: \"a04b62fe-2ebd-4b20-b4b0-93038ead2692\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.757168 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-nmstate-lock\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.757203 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-ovs-socket\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.773743 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.774469 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.776498 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dzd4c" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.776573 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.780437 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.788423 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.858916 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79986\" (UniqueName: \"kubernetes.io/projected/b5c62b83-eae8-455e-9c6e-166a9810c3f3-kube-api-access-79986\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859263 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhzvn\" (UniqueName: \"kubernetes.io/projected/11e4ce92-8234-4e31-a300-1896267b083c-kube-api-access-dhzvn\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8d62\" (UniqueName: \"kubernetes.io/projected/a04b62fe-2ebd-4b20-b4b0-93038ead2692-kube-api-access-k8d62\") pod \"nmstate-webhook-6b89b748d8-cbvck\" (UID: \"a04b62fe-2ebd-4b20-b4b0-93038ead2692\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859484 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzm89\" (UniqueName: \"kubernetes.io/projected/230299a9-48a6-405b-9569-33b7110107be-kube-api-access-pzm89\") pod \"nmstate-metrics-5dcf9c57c5-lbjdd\" (UID: \"230299a9-48a6-405b-9569-33b7110107be\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859600 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-dbus-socket\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859696 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b5c62b83-eae8-455e-9c6e-166a9810c3f3-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a04b62fe-2ebd-4b20-b4b0-93038ead2692-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-cbvck\" (UID: \"a04b62fe-2ebd-4b20-b4b0-93038ead2692\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-nmstate-lock\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.859983 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-nmstate-lock\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.860086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c62b83-eae8-455e-9c6e-166a9810c3f3-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.860182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-ovs-socket\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.860291 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-ovs-socket\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.860572 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/11e4ce92-8234-4e31-a300-1896267b083c-dbus-socket\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.871925 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/a04b62fe-2ebd-4b20-b4b0-93038ead2692-tls-key-pair\") pod \"nmstate-webhook-6b89b748d8-cbvck\" (UID: \"a04b62fe-2ebd-4b20-b4b0-93038ead2692\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.875344 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzm89\" (UniqueName: \"kubernetes.io/projected/230299a9-48a6-405b-9569-33b7110107be-kube-api-access-pzm89\") pod \"nmstate-metrics-5dcf9c57c5-lbjdd\" (UID: \"230299a9-48a6-405b-9569-33b7110107be\") " pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.876441 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhzvn\" (UniqueName: \"kubernetes.io/projected/11e4ce92-8234-4e31-a300-1896267b083c-kube-api-access-dhzvn\") pod \"nmstate-handler-42kdg\" (UID: \"11e4ce92-8234-4e31-a300-1896267b083c\") " pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.881527 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8d62\" (UniqueName: \"kubernetes.io/projected/a04b62fe-2ebd-4b20-b4b0-93038ead2692-kube-api-access-k8d62\") pod \"nmstate-webhook-6b89b748d8-cbvck\" (UID: \"a04b62fe-2ebd-4b20-b4b0-93038ead2692\") " pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.936770 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.952272 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.961020 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c62b83-eae8-455e-9c6e-166a9810c3f3-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.961101 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79986\" (UniqueName: \"kubernetes.io/projected/b5c62b83-eae8-455e-9c6e-166a9810c3f3-kube-api-access-79986\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.961164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b5c62b83-eae8-455e-9c6e-166a9810c3f3-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.962578 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/b5c62b83-eae8-455e-9c6e-166a9810c3f3-nginx-conf\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: E1125 10:00:33.962697 4926 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Nov 25 10:00:33 crc kubenswrapper[4926]: E1125 10:00:33.962779 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b5c62b83-eae8-455e-9c6e-166a9810c3f3-plugin-serving-cert podName:b5c62b83-eae8-455e-9c6e-166a9810c3f3 nodeName:}" failed. No retries permitted until 2025-11-25 10:00:34.462757865 +0000 UTC m=+731.146714292 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/b5c62b83-eae8-455e-9c6e-166a9810c3f3-plugin-serving-cert") pod "nmstate-console-plugin-5874bd7bc5-t42j4" (UID: "b5c62b83-eae8-455e-9c6e-166a9810c3f3") : secret "plugin-serving-cert" not found Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.971818 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-58f94ddcc-4vzhp"] Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.972565 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.976158 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.984741 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79986\" (UniqueName: \"kubernetes.io/projected/b5c62b83-eae8-455e-9c6e-166a9810c3f3-kube-api-access-79986\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:33 crc kubenswrapper[4926]: I1125 10:00:33.991842 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58f94ddcc-4vzhp"] Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062190 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-config\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062520 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-service-ca\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062561 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-oauth-config\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062579 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-oauth-serving-cert\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062599 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-trusted-ca-bundle\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062622 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhsb5\" (UniqueName: \"kubernetes.io/projected/55dcd2ee-7330-4f41-8c63-ed42bbad401b-kube-api-access-vhsb5\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.062654 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-serving-cert\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163403 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-service-ca\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163473 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-oauth-config\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163495 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-oauth-serving-cert\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163515 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-trusted-ca-bundle\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163537 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhsb5\" (UniqueName: \"kubernetes.io/projected/55dcd2ee-7330-4f41-8c63-ed42bbad401b-kube-api-access-vhsb5\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163576 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-serving-cert\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.163598 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-config\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.165044 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-config\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.165133 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-oauth-serving-cert\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.165163 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-service-ca\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.165157 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/55dcd2ee-7330-4f41-8c63-ed42bbad401b-trusted-ca-bundle\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.168307 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-serving-cert\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.173254 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/55dcd2ee-7330-4f41-8c63-ed42bbad401b-console-oauth-config\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.185634 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhsb5\" (UniqueName: \"kubernetes.io/projected/55dcd2ee-7330-4f41-8c63-ed42bbad401b-kube-api-access-vhsb5\") pod \"console-58f94ddcc-4vzhp\" (UID: \"55dcd2ee-7330-4f41-8c63-ed42bbad401b\") " pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.213205 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck"] Nov 25 10:00:34 crc kubenswrapper[4926]: W1125 10:00:34.222824 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda04b62fe_2ebd_4b20_b4b0_93038ead2692.slice/crio-758168c9a24910acb5b52a4f0ea2a6cedde818077fda479039b5a0dbed61cc90 WatchSource:0}: Error finding container 758168c9a24910acb5b52a4f0ea2a6cedde818077fda479039b5a0dbed61cc90: Status 404 returned error can't find the container with id 758168c9a24910acb5b52a4f0ea2a6cedde818077fda479039b5a0dbed61cc90 Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.322896 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.462813 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd"] Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.466946 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c62b83-eae8-455e-9c6e-166a9810c3f3-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:34 crc kubenswrapper[4926]: W1125 10:00:34.471216 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod230299a9_48a6_405b_9569_33b7110107be.slice/crio-f0c61c049318c24ca922307a0b8e371e91b5252cabddbfc5353240cca3731f98 WatchSource:0}: Error finding container f0c61c049318c24ca922307a0b8e371e91b5252cabddbfc5353240cca3731f98: Status 404 returned error can't find the container with id f0c61c049318c24ca922307a0b8e371e91b5252cabddbfc5353240cca3731f98 Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.471319 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/b5c62b83-eae8-455e-9c6e-166a9810c3f3-plugin-serving-cert\") pod \"nmstate-console-plugin-5874bd7bc5-t42j4\" (UID: \"b5c62b83-eae8-455e-9c6e-166a9810c3f3\") " pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.531061 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-58f94ddcc-4vzhp"] Nov 25 10:00:34 crc kubenswrapper[4926]: W1125 10:00:34.542206 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55dcd2ee_7330_4f41_8c63_ed42bbad401b.slice/crio-0314110b8d2467b799650836a81b2eddc4c9b150b03ee7c8d0e40ff8fa6c043b WatchSource:0}: Error finding container 0314110b8d2467b799650836a81b2eddc4c9b150b03ee7c8d0e40ff8fa6c043b: Status 404 returned error can't find the container with id 0314110b8d2467b799650836a81b2eddc4c9b150b03ee7c8d0e40ff8fa6c043b Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.567243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" event={"ID":"230299a9-48a6-405b-9569-33b7110107be","Type":"ContainerStarted","Data":"f0c61c049318c24ca922307a0b8e371e91b5252cabddbfc5353240cca3731f98"} Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.570002 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" event={"ID":"a04b62fe-2ebd-4b20-b4b0-93038ead2692","Type":"ContainerStarted","Data":"758168c9a24910acb5b52a4f0ea2a6cedde818077fda479039b5a0dbed61cc90"} Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.571615 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58f94ddcc-4vzhp" event={"ID":"55dcd2ee-7330-4f41-8c63-ed42bbad401b","Type":"ContainerStarted","Data":"0314110b8d2467b799650836a81b2eddc4c9b150b03ee7c8d0e40ff8fa6c043b"} Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.572693 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-42kdg" event={"ID":"11e4ce92-8234-4e31-a300-1896267b083c","Type":"ContainerStarted","Data":"a23dd9df0670c04b9c5d4dd61f65d0572dae65b6211244983dc3477237d4fdbd"} Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.688000 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" Nov 25 10:00:34 crc kubenswrapper[4926]: I1125 10:00:34.901602 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4"] Nov 25 10:00:34 crc kubenswrapper[4926]: W1125 10:00:34.911540 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5c62b83_eae8_455e_9c6e_166a9810c3f3.slice/crio-ee06fec3545bfa219a6f2e5ddadf989053d9d6a7a19e2ce1ec67bd091053456a WatchSource:0}: Error finding container ee06fec3545bfa219a6f2e5ddadf989053d9d6a7a19e2ce1ec67bd091053456a: Status 404 returned error can't find the container with id ee06fec3545bfa219a6f2e5ddadf989053d9d6a7a19e2ce1ec67bd091053456a Nov 25 10:00:35 crc kubenswrapper[4926]: I1125 10:00:35.593023 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" event={"ID":"b5c62b83-eae8-455e-9c6e-166a9810c3f3","Type":"ContainerStarted","Data":"ee06fec3545bfa219a6f2e5ddadf989053d9d6a7a19e2ce1ec67bd091053456a"} Nov 25 10:00:35 crc kubenswrapper[4926]: I1125 10:00:35.595380 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-58f94ddcc-4vzhp" event={"ID":"55dcd2ee-7330-4f41-8c63-ed42bbad401b","Type":"ContainerStarted","Data":"44b1232cf63861941e8c2cf31fbe957e51636c354dc09153a5f6c12f4187dc68"} Nov 25 10:00:35 crc kubenswrapper[4926]: I1125 10:00:35.622998 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-58f94ddcc-4vzhp" podStartSLOduration=2.6229607809999997 podStartE2EDuration="2.622960781s" podCreationTimestamp="2025-11-25 10:00:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:00:35.620725367 +0000 UTC m=+732.304681804" watchObservedRunningTime="2025-11-25 10:00:35.622960781 +0000 UTC m=+732.306917208" Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.615470 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" event={"ID":"a04b62fe-2ebd-4b20-b4b0-93038ead2692","Type":"ContainerStarted","Data":"cd572d3d5ee0d7777bcb9b30dab26b0202d053764cf6a739c903c37d649fc1eb"} Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.616449 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.618280 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-42kdg" event={"ID":"11e4ce92-8234-4e31-a300-1896267b083c","Type":"ContainerStarted","Data":"929c14a9aab91bdf7f299344b0042485edb8326acb4a7dc06b59213c5439789b"} Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.618453 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.619821 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" event={"ID":"230299a9-48a6-405b-9569-33b7110107be","Type":"ContainerStarted","Data":"b8e8dae8a4b3b33b5bc3af5baf31a06e9fb15c3166dde5579acd34179f6a4022"} Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.642009 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" podStartSLOduration=2.278884243 podStartE2EDuration="4.641984687s" podCreationTimestamp="2025-11-25 10:00:33 +0000 UTC" firstStartedPulling="2025-11-25 10:00:34.226957728 +0000 UTC m=+730.910914155" lastFinishedPulling="2025-11-25 10:00:36.590058142 +0000 UTC m=+733.274014599" observedRunningTime="2025-11-25 10:00:37.640286466 +0000 UTC m=+734.324242893" watchObservedRunningTime="2025-11-25 10:00:37.641984687 +0000 UTC m=+734.325941114" Nov 25 10:00:37 crc kubenswrapper[4926]: I1125 10:00:37.662305 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-42kdg" podStartSLOduration=2.203128021 podStartE2EDuration="4.662282144s" podCreationTimestamp="2025-11-25 10:00:33 +0000 UTC" firstStartedPulling="2025-11-25 10:00:34.027593841 +0000 UTC m=+730.711550268" lastFinishedPulling="2025-11-25 10:00:36.486747954 +0000 UTC m=+733.170704391" observedRunningTime="2025-11-25 10:00:37.660532151 +0000 UTC m=+734.344488588" watchObservedRunningTime="2025-11-25 10:00:37.662282144 +0000 UTC m=+734.346238571" Nov 25 10:00:38 crc kubenswrapper[4926]: I1125 10:00:38.630066 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" event={"ID":"b5c62b83-eae8-455e-9c6e-166a9810c3f3","Type":"ContainerStarted","Data":"d90a8e0c0193e2895bb2d7ee2e3cdda5502c54441aa42f38ae563dab0d4dc886"} Nov 25 10:00:38 crc kubenswrapper[4926]: I1125 10:00:38.647356 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-5874bd7bc5-t42j4" podStartSLOduration=2.809635868 podStartE2EDuration="5.647339062s" podCreationTimestamp="2025-11-25 10:00:33 +0000 UTC" firstStartedPulling="2025-11-25 10:00:34.916356274 +0000 UTC m=+731.600312701" lastFinishedPulling="2025-11-25 10:00:37.754059468 +0000 UTC m=+734.438015895" observedRunningTime="2025-11-25 10:00:38.643337974 +0000 UTC m=+735.327294401" watchObservedRunningTime="2025-11-25 10:00:38.647339062 +0000 UTC m=+735.331295489" Nov 25 10:00:39 crc kubenswrapper[4926]: I1125 10:00:39.641157 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" event={"ID":"230299a9-48a6-405b-9569-33b7110107be","Type":"ContainerStarted","Data":"d3c119a8da13a9d7eb7464c711bf5c0e851908f1e06146feffc542f1f1723b02"} Nov 25 10:00:39 crc kubenswrapper[4926]: I1125 10:00:39.677795 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-5dcf9c57c5-lbjdd" podStartSLOduration=2.313819848 podStartE2EDuration="6.677760222s" podCreationTimestamp="2025-11-25 10:00:33 +0000 UTC" firstStartedPulling="2025-11-25 10:00:34.474369861 +0000 UTC m=+731.158326288" lastFinishedPulling="2025-11-25 10:00:38.838310235 +0000 UTC m=+735.522266662" observedRunningTime="2025-11-25 10:00:39.669188742 +0000 UTC m=+736.353145209" watchObservedRunningTime="2025-11-25 10:00:39.677760222 +0000 UTC m=+736.361716679" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.008182 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-42kdg" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.015288 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.015352 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.323121 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.323185 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.332171 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.679414 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-58f94ddcc-4vzhp" Nov 25 10:00:44 crc kubenswrapper[4926]: I1125 10:00:44.749377 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-hfbwq"] Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.326613 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sw9lq"] Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.328728 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" podUID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" containerName="controller-manager" containerID="cri-o://2bb302c8a404f775fcb70e4691be106fb06de8cac1447c674f0fae23430cc2dc" gracePeriod=30 Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.428490 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd"] Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.428794 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" podUID="ceefca77-376e-4c60-8445-516d1d197f77" containerName="route-controller-manager" containerID="cri-o://38ca9fda52a4f0fa63c3018975ef9de273de601b95686354ba814b27444807c4" gracePeriod=30 Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.717249 4926 generic.go:334] "Generic (PLEG): container finished" podID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" containerID="2bb302c8a404f775fcb70e4691be106fb06de8cac1447c674f0fae23430cc2dc" exitCode=0 Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.717326 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" event={"ID":"65f3051c-91c1-44bc-a990-4f0a1165b4d2","Type":"ContainerDied","Data":"2bb302c8a404f775fcb70e4691be106fb06de8cac1447c674f0fae23430cc2dc"} Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.722049 4926 generic.go:334] "Generic (PLEG): container finished" podID="ceefca77-376e-4c60-8445-516d1d197f77" containerID="38ca9fda52a4f0fa63c3018975ef9de273de601b95686354ba814b27444807c4" exitCode=0 Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.722096 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" event={"ID":"ceefca77-376e-4c60-8445-516d1d197f77","Type":"ContainerDied","Data":"38ca9fda52a4f0fa63c3018975ef9de273de601b95686354ba814b27444807c4"} Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.767302 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.824728 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-config\") pod \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.824789 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-client-ca\") pod \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.824891 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-proxy-ca-bundles\") pod \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.824921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f3051c-91c1-44bc-a990-4f0a1165b4d2-serving-cert\") pod \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.824969 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7x6j\" (UniqueName: \"kubernetes.io/projected/65f3051c-91c1-44bc-a990-4f0a1165b4d2-kube-api-access-z7x6j\") pod \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\" (UID: \"65f3051c-91c1-44bc-a990-4f0a1165b4d2\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.825983 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "65f3051c-91c1-44bc-a990-4f0a1165b4d2" (UID: "65f3051c-91c1-44bc-a990-4f0a1165b4d2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.826120 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-config" (OuterVolumeSpecName: "config") pod "65f3051c-91c1-44bc-a990-4f0a1165b4d2" (UID: "65f3051c-91c1-44bc-a990-4f0a1165b4d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.826199 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-client-ca" (OuterVolumeSpecName: "client-ca") pod "65f3051c-91c1-44bc-a990-4f0a1165b4d2" (UID: "65f3051c-91c1-44bc-a990-4f0a1165b4d2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.833846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65f3051c-91c1-44bc-a990-4f0a1165b4d2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "65f3051c-91c1-44bc-a990-4f0a1165b4d2" (UID: "65f3051c-91c1-44bc-a990-4f0a1165b4d2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.833831 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.835905 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65f3051c-91c1-44bc-a990-4f0a1165b4d2-kube-api-access-z7x6j" (OuterVolumeSpecName: "kube-api-access-z7x6j") pod "65f3051c-91c1-44bc-a990-4f0a1165b4d2" (UID: "65f3051c-91c1-44bc-a990-4f0a1165b4d2"). InnerVolumeSpecName "kube-api-access-z7x6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.925880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-config\") pod \"ceefca77-376e-4c60-8445-516d1d197f77\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.925966 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xgkv\" (UniqueName: \"kubernetes.io/projected/ceefca77-376e-4c60-8445-516d1d197f77-kube-api-access-5xgkv\") pod \"ceefca77-376e-4c60-8445-516d1d197f77\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926010 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceefca77-376e-4c60-8445-516d1d197f77-serving-cert\") pod \"ceefca77-376e-4c60-8445-516d1d197f77\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926093 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-client-ca\") pod \"ceefca77-376e-4c60-8445-516d1d197f77\" (UID: \"ceefca77-376e-4c60-8445-516d1d197f77\") " Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926394 4926 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926423 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/65f3051c-91c1-44bc-a990-4f0a1165b4d2-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926435 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7x6j\" (UniqueName: \"kubernetes.io/projected/65f3051c-91c1-44bc-a990-4f0a1165b4d2-kube-api-access-z7x6j\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926452 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.926464 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/65f3051c-91c1-44bc-a990-4f0a1165b4d2-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.927463 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-config" (OuterVolumeSpecName: "config") pod "ceefca77-376e-4c60-8445-516d1d197f77" (UID: "ceefca77-376e-4c60-8445-516d1d197f77"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.927508 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-client-ca" (OuterVolumeSpecName: "client-ca") pod "ceefca77-376e-4c60-8445-516d1d197f77" (UID: "ceefca77-376e-4c60-8445-516d1d197f77"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.930144 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceefca77-376e-4c60-8445-516d1d197f77-kube-api-access-5xgkv" (OuterVolumeSpecName: "kube-api-access-5xgkv") pod "ceefca77-376e-4c60-8445-516d1d197f77" (UID: "ceefca77-376e-4c60-8445-516d1d197f77"). InnerVolumeSpecName "kube-api-access-5xgkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:00:49 crc kubenswrapper[4926]: I1125 10:00:49.930637 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceefca77-376e-4c60-8445-516d1d197f77-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ceefca77-376e-4c60-8445-516d1d197f77" (UID: "ceefca77-376e-4c60-8445-516d1d197f77"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.027448 4926 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-client-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.027738 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ceefca77-376e-4c60-8445-516d1d197f77-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.027816 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xgkv\" (UniqueName: \"kubernetes.io/projected/ceefca77-376e-4c60-8445-516d1d197f77-kube-api-access-5xgkv\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.027892 4926 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ceefca77-376e-4c60-8445-516d1d197f77-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.159226 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf"] Nov 25 10:00:50 crc kubenswrapper[4926]: E1125 10:00:50.159639 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceefca77-376e-4c60-8445-516d1d197f77" containerName="route-controller-manager" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.159671 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceefca77-376e-4c60-8445-516d1d197f77" containerName="route-controller-manager" Nov 25 10:00:50 crc kubenswrapper[4926]: E1125 10:00:50.159701 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" containerName="controller-manager" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.159714 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" containerName="controller-manager" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.159925 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" containerName="controller-manager" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.159996 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceefca77-376e-4c60-8445-516d1d197f77" containerName="route-controller-manager" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.160685 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.168987 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.193552 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.194604 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.209601 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.231529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-config\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.231624 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-proxy-ca-bundles\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.231703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40cfc92e-55ad-48e5-b67b-f686e067de2b-serving-cert\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.231733 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwsxg\" (UniqueName: \"kubernetes.io/projected/40cfc92e-55ad-48e5-b67b-f686e067de2b-kube-api-access-cwsxg\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.231957 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-client-ca\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.333696 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwsxg\" (UniqueName: \"kubernetes.io/projected/40cfc92e-55ad-48e5-b67b-f686e067de2b-kube-api-access-cwsxg\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.333765 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slq5k\" (UniqueName: \"kubernetes.io/projected/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-kube-api-access-slq5k\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.333795 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-client-ca\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.333831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-client-ca\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.333853 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-config\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.334069 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-config\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.334184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-proxy-ca-bundles\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.334446 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-serving-cert\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.334583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40cfc92e-55ad-48e5-b67b-f686e067de2b-serving-cert\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.334826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-client-ca\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.335849 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-config\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.336029 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/40cfc92e-55ad-48e5-b67b-f686e067de2b-proxy-ca-bundles\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.341730 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/40cfc92e-55ad-48e5-b67b-f686e067de2b-serving-cert\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.356679 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwsxg\" (UniqueName: \"kubernetes.io/projected/40cfc92e-55ad-48e5-b67b-f686e067de2b-kube-api-access-cwsxg\") pod \"controller-manager-5dbcd7449b-jtwvf\" (UID: \"40cfc92e-55ad-48e5-b67b-f686e067de2b\") " pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.435965 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-serving-cert\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.436085 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slq5k\" (UniqueName: \"kubernetes.io/projected/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-kube-api-access-slq5k\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.436135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-client-ca\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.436179 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-config\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.437790 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-client-ca\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.438637 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-config\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.448031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-serving-cert\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.459606 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slq5k\" (UniqueName: \"kubernetes.io/projected/b1a2a454-3cf9-4803-a54b-1eaf73e39fb2-kube-api-access-slq5k\") pod \"route-controller-manager-9494d7556-5bdck\" (UID: \"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2\") " pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.479438 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.509200 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.733324 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" event={"ID":"65f3051c-91c1-44bc-a990-4f0a1165b4d2","Type":"ContainerDied","Data":"5e8f0d0f74785d5347ffb8cc328e514a8019818efa93e4e9c00c1cf06301044b"} Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.734902 4926 scope.go:117] "RemoveContainer" containerID="2bb302c8a404f775fcb70e4691be106fb06de8cac1447c674f0fae23430cc2dc" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.733577 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-sw9lq" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.737305 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" event={"ID":"ceefca77-376e-4c60-8445-516d1d197f77","Type":"ContainerDied","Data":"15f74ecb5d6ce2364a9188c728b963caf0a9c2ba56fe3d40de30adfb0587bfb4"} Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.737394 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.771088 4926 scope.go:117] "RemoveContainer" containerID="38ca9fda52a4f0fa63c3018975ef9de273de601b95686354ba814b27444807c4" Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.782741 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sw9lq"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.787215 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-sw9lq"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.793085 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.796163 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-dw7jd"] Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.798996 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck"] Nov 25 10:00:50 crc kubenswrapper[4926]: W1125 10:00:50.803499 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1a2a454_3cf9_4803_a54b_1eaf73e39fb2.slice/crio-19a59a78aff7f215cd8618d9d62f4f9a1509e35060fb7f20c98bd9ea3aa679c9 WatchSource:0}: Error finding container 19a59a78aff7f215cd8618d9d62f4f9a1509e35060fb7f20c98bd9ea3aa679c9: Status 404 returned error can't find the container with id 19a59a78aff7f215cd8618d9d62f4f9a1509e35060fb7f20c98bd9ea3aa679c9 Nov 25 10:00:50 crc kubenswrapper[4926]: I1125 10:00:50.952204 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf"] Nov 25 10:00:50 crc kubenswrapper[4926]: W1125 10:00:50.960649 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod40cfc92e_55ad_48e5_b67b_f686e067de2b.slice/crio-532350e2ff8511dd89fdc4710aed80741727de51d59347bc63f4d3b69531d879 WatchSource:0}: Error finding container 532350e2ff8511dd89fdc4710aed80741727de51d59347bc63f4d3b69531d879: Status 404 returned error can't find the container with id 532350e2ff8511dd89fdc4710aed80741727de51d59347bc63f4d3b69531d879 Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.745255 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" event={"ID":"40cfc92e-55ad-48e5-b67b-f686e067de2b","Type":"ContainerStarted","Data":"ca745f7ad4675e2ae09b2d3dce6be3965e711a0e6bd586b116f9e86542b8591b"} Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.745664 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" event={"ID":"40cfc92e-55ad-48e5-b67b-f686e067de2b","Type":"ContainerStarted","Data":"532350e2ff8511dd89fdc4710aed80741727de51d59347bc63f4d3b69531d879"} Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.746283 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.749350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" event={"ID":"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2","Type":"ContainerStarted","Data":"51c533d846bc03ec19ade435a1e3c468fdb065d0cb9d717ddf35b3b62643c3db"} Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.749386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" event={"ID":"b1a2a454-3cf9-4803-a54b-1eaf73e39fb2","Type":"ContainerStarted","Data":"19a59a78aff7f215cd8618d9d62f4f9a1509e35060fb7f20c98bd9ea3aa679c9"} Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.749583 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.760445 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.761425 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.769361 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5dbcd7449b-jtwvf" podStartSLOduration=1.7693355830000002 podStartE2EDuration="1.769335583s" podCreationTimestamp="2025-11-25 10:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:00:51.768147623 +0000 UTC m=+748.452104050" watchObservedRunningTime="2025-11-25 10:00:51.769335583 +0000 UTC m=+748.453292010" Nov 25 10:00:51 crc kubenswrapper[4926]: I1125 10:00:51.789434 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-9494d7556-5bdck" podStartSLOduration=1.789405704 podStartE2EDuration="1.789405704s" podCreationTimestamp="2025-11-25 10:00:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:00:51.786157144 +0000 UTC m=+748.470113571" watchObservedRunningTime="2025-11-25 10:00:51.789405704 +0000 UTC m=+748.473362131" Nov 25 10:00:52 crc kubenswrapper[4926]: I1125 10:00:52.020106 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65f3051c-91c1-44bc-a990-4f0a1165b4d2" path="/var/lib/kubelet/pods/65f3051c-91c1-44bc-a990-4f0a1165b4d2/volumes" Nov 25 10:00:52 crc kubenswrapper[4926]: I1125 10:00:52.021101 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceefca77-376e-4c60-8445-516d1d197f77" path="/var/lib/kubelet/pods/ceefca77-376e-4c60-8445-516d1d197f77/volumes" Nov 25 10:00:53 crc kubenswrapper[4926]: I1125 10:00:53.960347 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6b89b748d8-cbvck" Nov 25 10:00:55 crc kubenswrapper[4926]: I1125 10:00:55.691118 4926 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.415667 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk"] Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.417908 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.421088 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.429534 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk"] Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.559750 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.559818 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbfgj\" (UniqueName: \"kubernetes.io/projected/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-kube-api-access-lbfgj\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.559871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.661612 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.661686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbfgj\" (UniqueName: \"kubernetes.io/projected/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-kube-api-access-lbfgj\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.661728 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.662360 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.662581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.688687 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbfgj\" (UniqueName: \"kubernetes.io/projected/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-kube-api-access-lbfgj\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.738540 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:09 crc kubenswrapper[4926]: I1125 10:01:09.800340 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-hfbwq" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerName="console" containerID="cri-o://5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845" gracePeriod=15 Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.203283 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk"] Nov 25 10:01:10 crc kubenswrapper[4926]: W1125 10:01:10.226912 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd69e0c8d_2117_461f_b01a_25e7bafb9ea2.slice/crio-ed588048fa74363e450f02328c32abd11fc39379a60c48a9138eeda03f283123 WatchSource:0}: Error finding container ed588048fa74363e450f02328c32abd11fc39379a60c48a9138eeda03f283123: Status 404 returned error can't find the container with id ed588048fa74363e450f02328c32abd11fc39379a60c48a9138eeda03f283123 Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.285849 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-hfbwq_60183186-5dd7-4632-96ef-2e87fb0c821a/console/0.log" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.286451 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.373563 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-serving-cert\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.373789 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-trusted-ca-bundle\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.373811 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-oauth-serving-cert\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.373948 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-oauth-config\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.374010 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-service-ca\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.374029 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-console-config\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.374082 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltmhq\" (UniqueName: \"kubernetes.io/projected/60183186-5dd7-4632-96ef-2e87fb0c821a-kube-api-access-ltmhq\") pod \"60183186-5dd7-4632-96ef-2e87fb0c821a\" (UID: \"60183186-5dd7-4632-96ef-2e87fb0c821a\") " Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.374954 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-console-config" (OuterVolumeSpecName: "console-config") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.374979 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.375365 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.375395 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-service-ca" (OuterVolumeSpecName: "service-ca") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.380189 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60183186-5dd7-4632-96ef-2e87fb0c821a-kube-api-access-ltmhq" (OuterVolumeSpecName: "kube-api-access-ltmhq") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "kube-api-access-ltmhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.380472 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.381247 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "60183186-5dd7-4632-96ef-2e87fb0c821a" (UID: "60183186-5dd7-4632-96ef-2e87fb0c821a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476389 4926 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-service-ca\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476439 4926 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-console-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476453 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltmhq\" (UniqueName: \"kubernetes.io/projected/60183186-5dd7-4632-96ef-2e87fb0c821a-kube-api-access-ltmhq\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476468 4926 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476481 4926 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476491 4926 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/60183186-5dd7-4632-96ef-2e87fb0c821a-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.476499 4926 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/60183186-5dd7-4632-96ef-2e87fb0c821a-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.882862 4926 generic.go:334] "Generic (PLEG): container finished" podID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerID="a4593cc24bfe580f9ff46b1eff1737540a906f97a8d2708884ea2c1e768e4171" exitCode=0 Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.883265 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" event={"ID":"d69e0c8d-2117-461f-b01a-25e7bafb9ea2","Type":"ContainerDied","Data":"a4593cc24bfe580f9ff46b1eff1737540a906f97a8d2708884ea2c1e768e4171"} Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.883444 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" event={"ID":"d69e0c8d-2117-461f-b01a-25e7bafb9ea2","Type":"ContainerStarted","Data":"ed588048fa74363e450f02328c32abd11fc39379a60c48a9138eeda03f283123"} Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.886967 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-hfbwq_60183186-5dd7-4632-96ef-2e87fb0c821a/console/0.log" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.887240 4926 generic.go:334] "Generic (PLEG): container finished" podID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerID="5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845" exitCode=2 Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.887253 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-hfbwq" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.887265 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hfbwq" event={"ID":"60183186-5dd7-4632-96ef-2e87fb0c821a","Type":"ContainerDied","Data":"5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845"} Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.887636 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-hfbwq" event={"ID":"60183186-5dd7-4632-96ef-2e87fb0c821a","Type":"ContainerDied","Data":"5079f08021b1945fcd94cfdf67fd311c165be4bcb058dc23aed30c3a1da78604"} Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.887694 4926 scope.go:117] "RemoveContainer" containerID="5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.936354 4926 scope.go:117] "RemoveContainer" containerID="5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845" Nov 25 10:01:10 crc kubenswrapper[4926]: E1125 10:01:10.945544 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845\": container with ID starting with 5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845 not found: ID does not exist" containerID="5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.945627 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845"} err="failed to get container status \"5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845\": rpc error: code = NotFound desc = could not find container \"5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845\": container with ID starting with 5dd32caf5a31da9b1bc3ca240a70e4748f1be40059767d5c61c0786b6aa3a845 not found: ID does not exist" Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.948001 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-hfbwq"] Nov 25 10:01:10 crc kubenswrapper[4926]: I1125 10:01:10.955273 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-hfbwq"] Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.022250 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" path="/var/lib/kubelet/pods/60183186-5dd7-4632-96ef-2e87fb0c821a/volumes" Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.904891 4926 generic.go:334] "Generic (PLEG): container finished" podID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerID="c99dbab1f2ea5a7b7a00891b6d06c1a81b120ae3a5627a2cc7407c4dfb6d71dc" exitCode=0 Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.905008 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" event={"ID":"d69e0c8d-2117-461f-b01a-25e7bafb9ea2","Type":"ContainerDied","Data":"c99dbab1f2ea5a7b7a00891b6d06c1a81b120ae3a5627a2cc7407c4dfb6d71dc"} Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.968070 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-skljb"] Nov 25 10:01:12 crc kubenswrapper[4926]: E1125 10:01:12.968507 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerName="console" Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.968533 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerName="console" Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.968699 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="60183186-5dd7-4632-96ef-2e87fb0c821a" containerName="console" Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.970192 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:12 crc kubenswrapper[4926]: I1125 10:01:12.980527 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-skljb"] Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.112990 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-catalog-content\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.114624 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-utilities\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.114808 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scgt5\" (UniqueName: \"kubernetes.io/projected/f751b0a1-7089-42be-8070-ba3acbdedfd4-kube-api-access-scgt5\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.236868 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-catalog-content\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.237275 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-utilities\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.237410 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scgt5\" (UniqueName: \"kubernetes.io/projected/f751b0a1-7089-42be-8070-ba3acbdedfd4-kube-api-access-scgt5\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.237561 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-catalog-content\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.237921 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-utilities\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.261913 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scgt5\" (UniqueName: \"kubernetes.io/projected/f751b0a1-7089-42be-8070-ba3acbdedfd4-kube-api-access-scgt5\") pod \"redhat-operators-skljb\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.328858 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.768053 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-skljb"] Nov 25 10:01:13 crc kubenswrapper[4926]: W1125 10:01:13.778665 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf751b0a1_7089_42be_8070_ba3acbdedfd4.slice/crio-0634c60ea1f03126ca1044b57e65c9cdb48fa354c3a7c97f399941b70bb360a8 WatchSource:0}: Error finding container 0634c60ea1f03126ca1044b57e65c9cdb48fa354c3a7c97f399941b70bb360a8: Status 404 returned error can't find the container with id 0634c60ea1f03126ca1044b57e65c9cdb48fa354c3a7c97f399941b70bb360a8 Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.916681 4926 generic.go:334] "Generic (PLEG): container finished" podID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerID="97964e3922e9c773bcc40361d86560867e2f18384323f23a8e05f3de935ea865" exitCode=0 Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.916796 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" event={"ID":"d69e0c8d-2117-461f-b01a-25e7bafb9ea2","Type":"ContainerDied","Data":"97964e3922e9c773bcc40361d86560867e2f18384323f23a8e05f3de935ea865"} Nov 25 10:01:13 crc kubenswrapper[4926]: I1125 10:01:13.924073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skljb" event={"ID":"f751b0a1-7089-42be-8070-ba3acbdedfd4","Type":"ContainerStarted","Data":"0634c60ea1f03126ca1044b57e65c9cdb48fa354c3a7c97f399941b70bb360a8"} Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.018961 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.019052 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.024401 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.025611 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d32c8b7deb1af6488ea2af61857b46c8de27ff7dc9d3d581bdb4b8b8f28e8724"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.025697 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://d32c8b7deb1af6488ea2af61857b46c8de27ff7dc9d3d581bdb4b8b8f28e8724" gracePeriod=600 Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.935346 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="d32c8b7deb1af6488ea2af61857b46c8de27ff7dc9d3d581bdb4b8b8f28e8724" exitCode=0 Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.935420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"d32c8b7deb1af6488ea2af61857b46c8de27ff7dc9d3d581bdb4b8b8f28e8724"} Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.936258 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"9bc7b6ba06aa7d0baa2979a51d5458380a1834ddc41f7963299878af0ce453ec"} Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.936295 4926 scope.go:117] "RemoveContainer" containerID="cbade86616f5a9cb6f1541e747cc180bfbb8672001b5ece28519dd37aebd3ef3" Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.939604 4926 generic.go:334] "Generic (PLEG): container finished" podID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerID="1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121" exitCode=0 Nov 25 10:01:14 crc kubenswrapper[4926]: I1125 10:01:14.939774 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skljb" event={"ID":"f751b0a1-7089-42be-8070-ba3acbdedfd4","Type":"ContainerDied","Data":"1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121"} Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.324785 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.474746 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbfgj\" (UniqueName: \"kubernetes.io/projected/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-kube-api-access-lbfgj\") pod \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.474914 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-bundle\") pod \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.475139 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-util\") pod \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\" (UID: \"d69e0c8d-2117-461f-b01a-25e7bafb9ea2\") " Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.475899 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-bundle" (OuterVolumeSpecName: "bundle") pod "d69e0c8d-2117-461f-b01a-25e7bafb9ea2" (UID: "d69e0c8d-2117-461f-b01a-25e7bafb9ea2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.481642 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-kube-api-access-lbfgj" (OuterVolumeSpecName: "kube-api-access-lbfgj") pod "d69e0c8d-2117-461f-b01a-25e7bafb9ea2" (UID: "d69e0c8d-2117-461f-b01a-25e7bafb9ea2"). InnerVolumeSpecName "kube-api-access-lbfgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.578047 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbfgj\" (UniqueName: \"kubernetes.io/projected/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-kube-api-access-lbfgj\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.578136 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.788280 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-util" (OuterVolumeSpecName: "util") pod "d69e0c8d-2117-461f-b01a-25e7bafb9ea2" (UID: "d69e0c8d-2117-461f-b01a-25e7bafb9ea2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.881718 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d69e0c8d-2117-461f-b01a-25e7bafb9ea2-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.956385 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" event={"ID":"d69e0c8d-2117-461f-b01a-25e7bafb9ea2","Type":"ContainerDied","Data":"ed588048fa74363e450f02328c32abd11fc39379a60c48a9138eeda03f283123"} Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.956436 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed588048fa74363e450f02328c32abd11fc39379a60c48a9138eeda03f283123" Nov 25 10:01:15 crc kubenswrapper[4926]: I1125 10:01:15.956504 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c6pqkgk" Nov 25 10:01:16 crc kubenswrapper[4926]: I1125 10:01:16.965494 4926 generic.go:334] "Generic (PLEG): container finished" podID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerID="49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e" exitCode=0 Nov 25 10:01:16 crc kubenswrapper[4926]: I1125 10:01:16.965550 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skljb" event={"ID":"f751b0a1-7089-42be-8070-ba3acbdedfd4","Type":"ContainerDied","Data":"49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e"} Nov 25 10:01:17 crc kubenswrapper[4926]: I1125 10:01:17.989071 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skljb" event={"ID":"f751b0a1-7089-42be-8070-ba3acbdedfd4","Type":"ContainerStarted","Data":"a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d"} Nov 25 10:01:18 crc kubenswrapper[4926]: I1125 10:01:18.014090 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-skljb" podStartSLOduration=3.557276516 podStartE2EDuration="6.014067221s" podCreationTimestamp="2025-11-25 10:01:12 +0000 UTC" firstStartedPulling="2025-11-25 10:01:14.941772218 +0000 UTC m=+771.625728645" lastFinishedPulling="2025-11-25 10:01:17.398562923 +0000 UTC m=+774.082519350" observedRunningTime="2025-11-25 10:01:18.011993911 +0000 UTC m=+774.695950348" watchObservedRunningTime="2025-11-25 10:01:18.014067221 +0000 UTC m=+774.698023658" Nov 25 10:01:23 crc kubenswrapper[4926]: I1125 10:01:23.329645 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:23 crc kubenswrapper[4926]: I1125 10:01:23.330335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:23 crc kubenswrapper[4926]: I1125 10:01:23.378716 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:24 crc kubenswrapper[4926]: I1125 10:01:24.071241 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:26 crc kubenswrapper[4926]: I1125 10:01:26.959768 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-skljb"] Nov 25 10:01:26 crc kubenswrapper[4926]: I1125 10:01:26.960575 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-skljb" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="registry-server" containerID="cri-o://a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d" gracePeriod=2 Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.136550 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg"] Nov 25 10:01:27 crc kubenswrapper[4926]: E1125 10:01:27.136829 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="util" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.136843 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="util" Nov 25 10:01:27 crc kubenswrapper[4926]: E1125 10:01:27.136865 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="pull" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.136872 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="pull" Nov 25 10:01:27 crc kubenswrapper[4926]: E1125 10:01:27.136878 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="extract" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.136885 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="extract" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.137021 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d69e0c8d-2117-461f-b01a-25e7bafb9ea2" containerName="extract" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.137471 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.140298 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.140989 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-vwftx" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.141127 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.143286 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.143901 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.155394 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-webhook-cert\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.155455 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh65j\" (UniqueName: \"kubernetes.io/projected/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-kube-api-access-fh65j\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.155574 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-apiservice-cert\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.156555 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg"] Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.256154 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-webhook-cert\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.256201 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh65j\" (UniqueName: \"kubernetes.io/projected/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-kube-api-access-fh65j\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.256259 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-apiservice-cert\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.263618 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-webhook-cert\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.278328 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh65j\" (UniqueName: \"kubernetes.io/projected/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-kube-api-access-fh65j\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.284230 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0cb1f5f8-6fcb-4664-8fcc-74e416d63eef-apiservice-cert\") pod \"metallb-operator-controller-manager-65f5fc5bb5-r6kqg\" (UID: \"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef\") " pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.456859 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.565250 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s"] Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.566267 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.568118 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.568232 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.568441 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pp2zc" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.649197 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s"] Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.661649 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c5e2728-0650-42d8-b0e0-212e87536355-webhook-cert\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.661693 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c5e2728-0650-42d8-b0e0-212e87536355-apiservice-cert\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.661756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79pkj\" (UniqueName: \"kubernetes.io/projected/7c5e2728-0650-42d8-b0e0-212e87536355-kube-api-access-79pkj\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.763926 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c5e2728-0650-42d8-b0e0-212e87536355-webhook-cert\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.764015 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c5e2728-0650-42d8-b0e0-212e87536355-apiservice-cert\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.764094 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79pkj\" (UniqueName: \"kubernetes.io/projected/7c5e2728-0650-42d8-b0e0-212e87536355-kube-api-access-79pkj\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.770574 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7c5e2728-0650-42d8-b0e0-212e87536355-webhook-cert\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.771185 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7c5e2728-0650-42d8-b0e0-212e87536355-apiservice-cert\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.783985 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79pkj\" (UniqueName: \"kubernetes.io/projected/7c5e2728-0650-42d8-b0e0-212e87536355-kube-api-access-79pkj\") pod \"metallb-operator-webhook-server-6f64c7cdd5-r4h6s\" (UID: \"7c5e2728-0650-42d8-b0e0-212e87536355\") " pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:27 crc kubenswrapper[4926]: I1125 10:01:27.885093 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.024751 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg"] Nov 25 10:01:28 crc kubenswrapper[4926]: W1125 10:01:28.032973 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cb1f5f8_6fcb_4664_8fcc_74e416d63eef.slice/crio-874261d8749f9a3531afbede80d63d9d64e75bcfb757e01dcb129da9363d84d6 WatchSource:0}: Error finding container 874261d8749f9a3531afbede80d63d9d64e75bcfb757e01dcb129da9363d84d6: Status 404 returned error can't find the container with id 874261d8749f9a3531afbede80d63d9d64e75bcfb757e01dcb129da9363d84d6 Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.076239 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" event={"ID":"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef","Type":"ContainerStarted","Data":"874261d8749f9a3531afbede80d63d9d64e75bcfb757e01dcb129da9363d84d6"} Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.179014 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s"] Nov 25 10:01:28 crc kubenswrapper[4926]: W1125 10:01:28.200668 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c5e2728_0650_42d8_b0e0_212e87536355.slice/crio-4c30712838b4fa5ed37bf26df0be5c1e241522dcf8f30624395b485391763bef WatchSource:0}: Error finding container 4c30712838b4fa5ed37bf26df0be5c1e241522dcf8f30624395b485391763bef: Status 404 returned error can't find the container with id 4c30712838b4fa5ed37bf26df0be5c1e241522dcf8f30624395b485391763bef Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.775810 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.852818 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-catalog-content\") pod \"f751b0a1-7089-42be-8070-ba3acbdedfd4\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.852921 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scgt5\" (UniqueName: \"kubernetes.io/projected/f751b0a1-7089-42be-8070-ba3acbdedfd4-kube-api-access-scgt5\") pod \"f751b0a1-7089-42be-8070-ba3acbdedfd4\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.852977 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-utilities\") pod \"f751b0a1-7089-42be-8070-ba3acbdedfd4\" (UID: \"f751b0a1-7089-42be-8070-ba3acbdedfd4\") " Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.855644 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-utilities" (OuterVolumeSpecName: "utilities") pod "f751b0a1-7089-42be-8070-ba3acbdedfd4" (UID: "f751b0a1-7089-42be-8070-ba3acbdedfd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.871290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f751b0a1-7089-42be-8070-ba3acbdedfd4-kube-api-access-scgt5" (OuterVolumeSpecName: "kube-api-access-scgt5") pod "f751b0a1-7089-42be-8070-ba3acbdedfd4" (UID: "f751b0a1-7089-42be-8070-ba3acbdedfd4"). InnerVolumeSpecName "kube-api-access-scgt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.954587 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scgt5\" (UniqueName: \"kubernetes.io/projected/f751b0a1-7089-42be-8070-ba3acbdedfd4-kube-api-access-scgt5\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.954667 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:28 crc kubenswrapper[4926]: I1125 10:01:28.988053 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f751b0a1-7089-42be-8070-ba3acbdedfd4" (UID: "f751b0a1-7089-42be-8070-ba3acbdedfd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.056571 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f751b0a1-7089-42be-8070-ba3acbdedfd4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.084980 4926 generic.go:334] "Generic (PLEG): container finished" podID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerID="a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d" exitCode=0 Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.085085 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-skljb" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.085129 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skljb" event={"ID":"f751b0a1-7089-42be-8070-ba3acbdedfd4","Type":"ContainerDied","Data":"a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d"} Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.085164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-skljb" event={"ID":"f751b0a1-7089-42be-8070-ba3acbdedfd4","Type":"ContainerDied","Data":"0634c60ea1f03126ca1044b57e65c9cdb48fa354c3a7c97f399941b70bb360a8"} Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.085190 4926 scope.go:117] "RemoveContainer" containerID="a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.087738 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" event={"ID":"7c5e2728-0650-42d8-b0e0-212e87536355","Type":"ContainerStarted","Data":"4c30712838b4fa5ed37bf26df0be5c1e241522dcf8f30624395b485391763bef"} Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.111013 4926 scope.go:117] "RemoveContainer" containerID="49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.124197 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-skljb"] Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.132282 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-skljb"] Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.139637 4926 scope.go:117] "RemoveContainer" containerID="1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.168540 4926 scope.go:117] "RemoveContainer" containerID="a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d" Nov 25 10:01:29 crc kubenswrapper[4926]: E1125 10:01:29.170489 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d\": container with ID starting with a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d not found: ID does not exist" containerID="a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.170632 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d"} err="failed to get container status \"a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d\": rpc error: code = NotFound desc = could not find container \"a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d\": container with ID starting with a225b5002b7a25ec26441fd42623b4ab9d3209c49cc1070b75b58cc0ea9f641d not found: ID does not exist" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.170712 4926 scope.go:117] "RemoveContainer" containerID="49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e" Nov 25 10:01:29 crc kubenswrapper[4926]: E1125 10:01:29.174459 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e\": container with ID starting with 49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e not found: ID does not exist" containerID="49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.174548 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e"} err="failed to get container status \"49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e\": rpc error: code = NotFound desc = could not find container \"49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e\": container with ID starting with 49857492f97bc2981ed7dc6601ec80497b121caf5c8c6e1535336e8b11806c2e not found: ID does not exist" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.174623 4926 scope.go:117] "RemoveContainer" containerID="1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121" Nov 25 10:01:29 crc kubenswrapper[4926]: E1125 10:01:29.177743 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121\": container with ID starting with 1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121 not found: ID does not exist" containerID="1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121" Nov 25 10:01:29 crc kubenswrapper[4926]: I1125 10:01:29.177843 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121"} err="failed to get container status \"1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121\": rpc error: code = NotFound desc = could not find container \"1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121\": container with ID starting with 1869e7d0bb78e8b1cccb19a2f57078ded37f048057e97c5def4d1254ccabf121 not found: ID does not exist" Nov 25 10:01:30 crc kubenswrapper[4926]: I1125 10:01:30.023744 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" path="/var/lib/kubelet/pods/f751b0a1-7089-42be-8070-ba3acbdedfd4/volumes" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.769711 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7scxc"] Nov 25 10:01:31 crc kubenswrapper[4926]: E1125 10:01:31.770463 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="extract-content" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.770481 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="extract-content" Nov 25 10:01:31 crc kubenswrapper[4926]: E1125 10:01:31.770496 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="extract-utilities" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.770504 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="extract-utilities" Nov 25 10:01:31 crc kubenswrapper[4926]: E1125 10:01:31.770530 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="registry-server" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.770538 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="registry-server" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.770700 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f751b0a1-7089-42be-8070-ba3acbdedfd4" containerName="registry-server" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.771811 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.790621 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7scxc"] Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.905303 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-utilities\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.905371 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hqrh\" (UniqueName: \"kubernetes.io/projected/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-kube-api-access-2hqrh\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:31 crc kubenswrapper[4926]: I1125 10:01:31.905650 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-catalog-content\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.007541 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-catalog-content\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.007736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-utilities\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.007759 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hqrh\" (UniqueName: \"kubernetes.io/projected/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-kube-api-access-2hqrh\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.008341 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-catalog-content\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.008490 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-utilities\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.045309 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hqrh\" (UniqueName: \"kubernetes.io/projected/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-kube-api-access-2hqrh\") pod \"community-operators-7scxc\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.087335 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.120434 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" event={"ID":"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef","Type":"ContainerStarted","Data":"0580482c8e707e87bce39e86dd5a0fc30e705dc66256d88ecce3aab5b1b165e4"} Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.120661 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:01:32 crc kubenswrapper[4926]: I1125 10:01:32.145750 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" podStartSLOduration=1.9977402450000001 podStartE2EDuration="5.145721181s" podCreationTimestamp="2025-11-25 10:01:27 +0000 UTC" firstStartedPulling="2025-11-25 10:01:28.038206451 +0000 UTC m=+784.722162878" lastFinishedPulling="2025-11-25 10:01:31.186187387 +0000 UTC m=+787.870143814" observedRunningTime="2025-11-25 10:01:32.142925292 +0000 UTC m=+788.826881719" watchObservedRunningTime="2025-11-25 10:01:32.145721181 +0000 UTC m=+788.829677618" Nov 25 10:01:33 crc kubenswrapper[4926]: I1125 10:01:33.615121 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7scxc"] Nov 25 10:01:34 crc kubenswrapper[4926]: I1125 10:01:34.147333 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" event={"ID":"7c5e2728-0650-42d8-b0e0-212e87536355","Type":"ContainerStarted","Data":"0b94be2db10e9792fafd05370ee9dbb1e546cd06700060e26d5b5e55cd677aaa"} Nov 25 10:01:34 crc kubenswrapper[4926]: I1125 10:01:34.147879 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:34 crc kubenswrapper[4926]: I1125 10:01:34.150742 4926 generic.go:334] "Generic (PLEG): container finished" podID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerID="2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1" exitCode=0 Nov 25 10:01:34 crc kubenswrapper[4926]: I1125 10:01:34.150787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7scxc" event={"ID":"57dd6d19-b0ce-4923-8df4-9ce63e7bd811","Type":"ContainerDied","Data":"2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1"} Nov 25 10:01:34 crc kubenswrapper[4926]: I1125 10:01:34.150814 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7scxc" event={"ID":"57dd6d19-b0ce-4923-8df4-9ce63e7bd811","Type":"ContainerStarted","Data":"40e1f72a7feb30219995391b2fa23674a01eb5a119606798652444e589fbc2fb"} Nov 25 10:01:34 crc kubenswrapper[4926]: I1125 10:01:34.171206 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" podStartSLOduration=2.145192673 podStartE2EDuration="7.171183344s" podCreationTimestamp="2025-11-25 10:01:27 +0000 UTC" firstStartedPulling="2025-11-25 10:01:28.205577705 +0000 UTC m=+784.889534142" lastFinishedPulling="2025-11-25 10:01:33.231568386 +0000 UTC m=+789.915524813" observedRunningTime="2025-11-25 10:01:34.168492078 +0000 UTC m=+790.852448515" watchObservedRunningTime="2025-11-25 10:01:34.171183344 +0000 UTC m=+790.855139771" Nov 25 10:01:36 crc kubenswrapper[4926]: I1125 10:01:36.165272 4926 generic.go:334] "Generic (PLEG): container finished" podID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerID="8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27" exitCode=0 Nov 25 10:01:36 crc kubenswrapper[4926]: I1125 10:01:36.165333 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7scxc" event={"ID":"57dd6d19-b0ce-4923-8df4-9ce63e7bd811","Type":"ContainerDied","Data":"8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27"} Nov 25 10:01:37 crc kubenswrapper[4926]: I1125 10:01:37.176038 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7scxc" event={"ID":"57dd6d19-b0ce-4923-8df4-9ce63e7bd811","Type":"ContainerStarted","Data":"f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86"} Nov 25 10:01:37 crc kubenswrapper[4926]: I1125 10:01:37.205869 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7scxc" podStartSLOduration=3.746526301 podStartE2EDuration="6.205844287s" podCreationTimestamp="2025-11-25 10:01:31 +0000 UTC" firstStartedPulling="2025-11-25 10:01:34.155095741 +0000 UTC m=+790.839052158" lastFinishedPulling="2025-11-25 10:01:36.614413717 +0000 UTC m=+793.298370144" observedRunningTime="2025-11-25 10:01:37.202388822 +0000 UTC m=+793.886345249" watchObservedRunningTime="2025-11-25 10:01:37.205844287 +0000 UTC m=+793.889800714" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.165703 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-7ltkk"] Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.167487 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.177407 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ltkk"] Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.328019 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-utilities\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.328535 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-catalog-content\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.328605 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2qw6\" (UniqueName: \"kubernetes.io/projected/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-kube-api-access-t2qw6\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.430137 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-utilities\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.430519 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-catalog-content\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.430685 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2qw6\" (UniqueName: \"kubernetes.io/projected/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-kube-api-access-t2qw6\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.430771 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-utilities\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.431214 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-catalog-content\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.458556 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2qw6\" (UniqueName: \"kubernetes.io/projected/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-kube-api-access-t2qw6\") pod \"redhat-marketplace-7ltkk\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.487322 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:39 crc kubenswrapper[4926]: I1125 10:01:39.936612 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ltkk"] Nov 25 10:01:40 crc kubenswrapper[4926]: I1125 10:01:40.196481 4926 generic.go:334] "Generic (PLEG): container finished" podID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerID="527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0" exitCode=0 Nov 25 10:01:40 crc kubenswrapper[4926]: I1125 10:01:40.196558 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ltkk" event={"ID":"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de","Type":"ContainerDied","Data":"527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0"} Nov 25 10:01:40 crc kubenswrapper[4926]: I1125 10:01:40.196601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ltkk" event={"ID":"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de","Type":"ContainerStarted","Data":"c657cc460ab3cbced1e0bd83cc80aa6656d8b39570564de806f311a87fa692c1"} Nov 25 10:01:41 crc kubenswrapper[4926]: I1125 10:01:41.205068 4926 generic.go:334] "Generic (PLEG): container finished" podID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerID="9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9" exitCode=0 Nov 25 10:01:41 crc kubenswrapper[4926]: I1125 10:01:41.205116 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ltkk" event={"ID":"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de","Type":"ContainerDied","Data":"9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9"} Nov 25 10:01:42 crc kubenswrapper[4926]: I1125 10:01:42.088109 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:42 crc kubenswrapper[4926]: I1125 10:01:42.088632 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:42 crc kubenswrapper[4926]: I1125 10:01:42.134960 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:42 crc kubenswrapper[4926]: I1125 10:01:42.213900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ltkk" event={"ID":"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de","Type":"ContainerStarted","Data":"e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994"} Nov 25 10:01:42 crc kubenswrapper[4926]: I1125 10:01:42.238881 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-7ltkk" podStartSLOduration=1.8211065419999999 podStartE2EDuration="3.238863438s" podCreationTimestamp="2025-11-25 10:01:39 +0000 UTC" firstStartedPulling="2025-11-25 10:01:40.199498116 +0000 UTC m=+796.883454543" lastFinishedPulling="2025-11-25 10:01:41.617255012 +0000 UTC m=+798.301211439" observedRunningTime="2025-11-25 10:01:42.237369172 +0000 UTC m=+798.921325599" watchObservedRunningTime="2025-11-25 10:01:42.238863438 +0000 UTC m=+798.922819865" Nov 25 10:01:42 crc kubenswrapper[4926]: I1125 10:01:42.258864 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:45 crc kubenswrapper[4926]: I1125 10:01:45.754513 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7scxc"] Nov 25 10:01:45 crc kubenswrapper[4926]: I1125 10:01:45.755350 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-7scxc" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="registry-server" containerID="cri-o://f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86" gracePeriod=2 Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.193184 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.288242 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7scxc" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.288602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7scxc" event={"ID":"57dd6d19-b0ce-4923-8df4-9ce63e7bd811","Type":"ContainerDied","Data":"f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86"} Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.288666 4926 scope.go:117] "RemoveContainer" containerID="f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.290337 4926 generic.go:334] "Generic (PLEG): container finished" podID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerID="f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86" exitCode=0 Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.290387 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7scxc" event={"ID":"57dd6d19-b0ce-4923-8df4-9ce63e7bd811","Type":"ContainerDied","Data":"40e1f72a7feb30219995391b2fa23674a01eb5a119606798652444e589fbc2fb"} Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.313108 4926 scope.go:117] "RemoveContainer" containerID="8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.332860 4926 scope.go:117] "RemoveContainer" containerID="2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.351608 4926 scope.go:117] "RemoveContainer" containerID="f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86" Nov 25 10:01:46 crc kubenswrapper[4926]: E1125 10:01:46.352451 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86\": container with ID starting with f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86 not found: ID does not exist" containerID="f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.352539 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86"} err="failed to get container status \"f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86\": rpc error: code = NotFound desc = could not find container \"f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86\": container with ID starting with f5c17f362eeb2b991e2ef49dfd7381c3467e137d2be36e677de1f91f42eefe86 not found: ID does not exist" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.352564 4926 scope.go:117] "RemoveContainer" containerID="8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27" Nov 25 10:01:46 crc kubenswrapper[4926]: E1125 10:01:46.353607 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27\": container with ID starting with 8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27 not found: ID does not exist" containerID="8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.353631 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27"} err="failed to get container status \"8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27\": rpc error: code = NotFound desc = could not find container \"8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27\": container with ID starting with 8972993c7c6a2b0999c6d9cd27c9907806d5a219006c3bd26c82b30c853d6f27 not found: ID does not exist" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.353647 4926 scope.go:117] "RemoveContainer" containerID="2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1" Nov 25 10:01:46 crc kubenswrapper[4926]: E1125 10:01:46.354026 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1\": container with ID starting with 2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1 not found: ID does not exist" containerID="2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.354046 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1"} err="failed to get container status \"2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1\": rpc error: code = NotFound desc = could not find container \"2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1\": container with ID starting with 2eddd88a1b9c69c174b6988795f610f6e05918d6f52092b5cc5da985e4157bc1 not found: ID does not exist" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.366798 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-catalog-content\") pod \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.366964 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-utilities\") pod \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.367067 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hqrh\" (UniqueName: \"kubernetes.io/projected/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-kube-api-access-2hqrh\") pod \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\" (UID: \"57dd6d19-b0ce-4923-8df4-9ce63e7bd811\") " Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.368541 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-utilities" (OuterVolumeSpecName: "utilities") pod "57dd6d19-b0ce-4923-8df4-9ce63e7bd811" (UID: "57dd6d19-b0ce-4923-8df4-9ce63e7bd811"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.374556 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-kube-api-access-2hqrh" (OuterVolumeSpecName: "kube-api-access-2hqrh") pod "57dd6d19-b0ce-4923-8df4-9ce63e7bd811" (UID: "57dd6d19-b0ce-4923-8df4-9ce63e7bd811"). InnerVolumeSpecName "kube-api-access-2hqrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.448892 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57dd6d19-b0ce-4923-8df4-9ce63e7bd811" (UID: "57dd6d19-b0ce-4923-8df4-9ce63e7bd811"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.473146 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hqrh\" (UniqueName: \"kubernetes.io/projected/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-kube-api-access-2hqrh\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.473196 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.473207 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57dd6d19-b0ce-4923-8df4-9ce63e7bd811-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.623097 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-7scxc"] Nov 25 10:01:46 crc kubenswrapper[4926]: I1125 10:01:46.626483 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-7scxc"] Nov 25 10:01:47 crc kubenswrapper[4926]: I1125 10:01:47.896458 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6f64c7cdd5-r4h6s" Nov 25 10:01:48 crc kubenswrapper[4926]: I1125 10:01:48.020764 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" path="/var/lib/kubelet/pods/57dd6d19-b0ce-4923-8df4-9ce63e7bd811/volumes" Nov 25 10:01:49 crc kubenswrapper[4926]: I1125 10:01:49.488238 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:49 crc kubenswrapper[4926]: I1125 10:01:49.488356 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:49 crc kubenswrapper[4926]: I1125 10:01:49.535433 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:50 crc kubenswrapper[4926]: I1125 10:01:50.371427 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:52 crc kubenswrapper[4926]: I1125 10:01:52.555750 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ltkk"] Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.333478 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-7ltkk" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="registry-server" containerID="cri-o://e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994" gracePeriod=2 Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.752767 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.800158 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2qw6\" (UniqueName: \"kubernetes.io/projected/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-kube-api-access-t2qw6\") pod \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.800226 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-catalog-content\") pod \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.800289 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-utilities\") pod \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\" (UID: \"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de\") " Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.801539 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-utilities" (OuterVolumeSpecName: "utilities") pod "8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" (UID: "8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.806953 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-kube-api-access-t2qw6" (OuterVolumeSpecName: "kube-api-access-t2qw6") pod "8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" (UID: "8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de"). InnerVolumeSpecName "kube-api-access-t2qw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.819053 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" (UID: "8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.901742 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.901798 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2qw6\" (UniqueName: \"kubernetes.io/projected/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-kube-api-access-t2qw6\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:53 crc kubenswrapper[4926]: I1125 10:01:53.901815 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.342061 4926 generic.go:334] "Generic (PLEG): container finished" podID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerID="e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994" exitCode=0 Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.342136 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-7ltkk" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.342164 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ltkk" event={"ID":"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de","Type":"ContainerDied","Data":"e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994"} Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.342483 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-7ltkk" event={"ID":"8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de","Type":"ContainerDied","Data":"c657cc460ab3cbced1e0bd83cc80aa6656d8b39570564de806f311a87fa692c1"} Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.342508 4926 scope.go:117] "RemoveContainer" containerID="e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.364814 4926 scope.go:117] "RemoveContainer" containerID="9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.366773 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ltkk"] Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.374473 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-7ltkk"] Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.393460 4926 scope.go:117] "RemoveContainer" containerID="527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.419352 4926 scope.go:117] "RemoveContainer" containerID="e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994" Nov 25 10:01:54 crc kubenswrapper[4926]: E1125 10:01:54.419606 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994\": container with ID starting with e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994 not found: ID does not exist" containerID="e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.419647 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994"} err="failed to get container status \"e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994\": rpc error: code = NotFound desc = could not find container \"e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994\": container with ID starting with e5a9f33ed9c1d3020049aeec9a203d6222e87df1591bd955dd7e61b5dd436994 not found: ID does not exist" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.419680 4926 scope.go:117] "RemoveContainer" containerID="9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9" Nov 25 10:01:54 crc kubenswrapper[4926]: E1125 10:01:54.419872 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9\": container with ID starting with 9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9 not found: ID does not exist" containerID="9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.419893 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9"} err="failed to get container status \"9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9\": rpc error: code = NotFound desc = could not find container \"9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9\": container with ID starting with 9e349d9f053cf0daf8607f11994bc51955a6ed4042b5f4c4316565573cf3a4f9 not found: ID does not exist" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.419911 4926 scope.go:117] "RemoveContainer" containerID="527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0" Nov 25 10:01:54 crc kubenswrapper[4926]: E1125 10:01:54.420111 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0\": container with ID starting with 527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0 not found: ID does not exist" containerID="527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0" Nov 25 10:01:54 crc kubenswrapper[4926]: I1125 10:01:54.420134 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0"} err="failed to get container status \"527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0\": rpc error: code = NotFound desc = could not find container \"527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0\": container with ID starting with 527191608e8a80c305864f401927a2e8cc026d751f4318a387b1595e6c0bfbf0 not found: ID does not exist" Nov 25 10:01:56 crc kubenswrapper[4926]: I1125 10:01:56.021617 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" path="/var/lib/kubelet/pods/8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de/volumes" Nov 25 10:02:07 crc kubenswrapper[4926]: I1125 10:02:07.459971 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230137 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-49pz2"] Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.230447 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="registry-server" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230463 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="registry-server" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.230485 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="extract-content" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230492 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="extract-content" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.230503 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="extract-content" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230510 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="extract-content" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.230517 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="registry-server" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230523 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="registry-server" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.230537 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="extract-utilities" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230543 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="extract-utilities" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.230560 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="extract-utilities" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230566 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="extract-utilities" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230708 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="57dd6d19-b0ce-4923-8df4-9ce63e7bd811" containerName="registry-server" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.230717 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c8cb98a-d8bb-4e89-9340-e7d7cecfe4de" containerName="registry-server" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.232855 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.236234 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.236417 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6vlnp" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.236617 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.241884 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb"] Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.242917 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.250568 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.258146 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb"] Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.327167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-metrics-certs\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.327714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-reloader\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.327758 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c8zm\" (UniqueName: \"kubernetes.io/projected/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-kube-api-access-6c8zm\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.327793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-sockets\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.328017 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-metrics\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.328129 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2-cert\") pod \"frr-k8s-webhook-server-6998585d5-6ztkb\" (UID: \"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.328170 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-startup\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.328201 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsfkv\" (UniqueName: \"kubernetes.io/projected/1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2-kube-api-access-tsfkv\") pod \"frr-k8s-webhook-server-6998585d5-6ztkb\" (UID: \"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.328247 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-conf\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.344970 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-6rxct"] Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.346926 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.351389 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qqs4j" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.351501 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.351603 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.353520 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.367422 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-bvqpc"] Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.368278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.370557 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.388264 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-bvqpc"] Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.428990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2-cert\") pod \"frr-k8s-webhook-server-6998585d5-6ztkb\" (UID: \"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429038 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-startup\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429061 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsfkv\" (UniqueName: \"kubernetes.io/projected/1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2-kube-api-access-tsfkv\") pod \"frr-k8s-webhook-server-6998585d5-6ztkb\" (UID: \"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-conf\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429121 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-metrics-certs\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429150 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429173 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldcms\" (UniqueName: \"kubernetes.io/projected/7586af8a-a829-4c9f-9589-a3643c147f1d-kube-api-access-ldcms\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429195 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-reloader\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429213 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c8zm\" (UniqueName: \"kubernetes.io/projected/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-kube-api-access-6c8zm\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-sockets\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429254 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-metrics-certs\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429289 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-metrics\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.429305 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7586af8a-a829-4c9f-9589-a3643c147f1d-metallb-excludel2\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.430414 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-reloader\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.430524 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-startup\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.430559 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-conf\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.430683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-frr-sockets\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.431174 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-metrics\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.437290 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2-cert\") pod \"frr-k8s-webhook-server-6998585d5-6ztkb\" (UID: \"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.438327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-metrics-certs\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.454766 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsfkv\" (UniqueName: \"kubernetes.io/projected/1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2-kube-api-access-tsfkv\") pod \"frr-k8s-webhook-server-6998585d5-6ztkb\" (UID: \"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.458836 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c8zm\" (UniqueName: \"kubernetes.io/projected/99b2fe2e-14bb-4dda-9152-bb593b6cb80e-kube-api-access-6c8zm\") pod \"frr-k8s-49pz2\" (UID: \"99b2fe2e-14bb-4dda-9152-bb593b6cb80e\") " pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531306 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765233dc-e1b5-47b8-8eb5-8cef0a62f428-cert\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531375 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765233dc-e1b5-47b8-8eb5-8cef0a62f428-metrics-certs\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531399 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thcss\" (UniqueName: \"kubernetes.io/projected/765233dc-e1b5-47b8-8eb5-8cef0a62f428-kube-api-access-thcss\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531433 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531459 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldcms\" (UniqueName: \"kubernetes.io/projected/7586af8a-a829-4c9f-9589-a3643c147f1d-kube-api-access-ldcms\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.531654 4926 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531694 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-metrics-certs\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: E1125 10:02:08.531790 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist podName:7586af8a-a829-4c9f-9589-a3643c147f1d nodeName:}" failed. No retries permitted until 2025-11-25 10:02:09.031755754 +0000 UTC m=+825.715712191 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist") pod "speaker-6rxct" (UID: "7586af8a-a829-4c9f-9589-a3643c147f1d") : secret "metallb-memberlist" not found Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.531895 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7586af8a-a829-4c9f-9589-a3643c147f1d-metallb-excludel2\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.532810 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7586af8a-a829-4c9f-9589-a3643c147f1d-metallb-excludel2\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.535423 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-metrics-certs\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.549236 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldcms\" (UniqueName: \"kubernetes.io/projected/7586af8a-a829-4c9f-9589-a3643c147f1d-kube-api-access-ldcms\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.570619 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.588243 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.633235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765233dc-e1b5-47b8-8eb5-8cef0a62f428-metrics-certs\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.633290 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thcss\" (UniqueName: \"kubernetes.io/projected/765233dc-e1b5-47b8-8eb5-8cef0a62f428-kube-api-access-thcss\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.633312 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765233dc-e1b5-47b8-8eb5-8cef0a62f428-cert\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.639974 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/765233dc-e1b5-47b8-8eb5-8cef0a62f428-cert\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.640172 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/765233dc-e1b5-47b8-8eb5-8cef0a62f428-metrics-certs\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.652982 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thcss\" (UniqueName: \"kubernetes.io/projected/765233dc-e1b5-47b8-8eb5-8cef0a62f428-kube-api-access-thcss\") pod \"controller-6c7b4b5f48-bvqpc\" (UID: \"765233dc-e1b5-47b8-8eb5-8cef0a62f428\") " pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.688193 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:08 crc kubenswrapper[4926]: I1125 10:02:08.901506 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-bvqpc"] Nov 25 10:02:08 crc kubenswrapper[4926]: W1125 10:02:08.906715 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod765233dc_e1b5_47b8_8eb5_8cef0a62f428.slice/crio-1fdd6c31d11dd5f6f2d1de0eb9cb22fa38e74c4af1e4da35b278d3c2fc1ba7c5 WatchSource:0}: Error finding container 1fdd6c31d11dd5f6f2d1de0eb9cb22fa38e74c4af1e4da35b278d3c2fc1ba7c5: Status 404 returned error can't find the container with id 1fdd6c31d11dd5f6f2d1de0eb9cb22fa38e74c4af1e4da35b278d3c2fc1ba7c5 Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.020952 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb"] Nov 25 10:02:09 crc kubenswrapper[4926]: W1125 10:02:09.027011 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a8ae06b_bbc4_4854_9ebd_a88bf4c7cde2.slice/crio-4cec178adf4c9faaadaee658d75df0bdeb7062e48673efc1991a7ee3639c642a WatchSource:0}: Error finding container 4cec178adf4c9faaadaee658d75df0bdeb7062e48673efc1991a7ee3639c642a: Status 404 returned error can't find the container with id 4cec178adf4c9faaadaee658d75df0bdeb7062e48673efc1991a7ee3639c642a Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.047997 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:09 crc kubenswrapper[4926]: E1125 10:02:09.048169 4926 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 25 10:02:09 crc kubenswrapper[4926]: E1125 10:02:09.048235 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist podName:7586af8a-a829-4c9f-9589-a3643c147f1d nodeName:}" failed. No retries permitted until 2025-11-25 10:02:10.048214469 +0000 UTC m=+826.732170896 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist") pod "speaker-6rxct" (UID: "7586af8a-a829-4c9f-9589-a3643c147f1d") : secret "metallb-memberlist" not found Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.447011 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-bvqpc" event={"ID":"765233dc-e1b5-47b8-8eb5-8cef0a62f428","Type":"ContainerStarted","Data":"2907e1daea4a252d3dfba2c0f4a0cb9cfc8a33d6133d15d5a8403c7cdf2f508b"} Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.447535 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-bvqpc" event={"ID":"765233dc-e1b5-47b8-8eb5-8cef0a62f428","Type":"ContainerStarted","Data":"c13b831f48deccd795860a4e8c8de45823dc328ea9d42a88c71a03e4faddc71e"} Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.447554 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-bvqpc" event={"ID":"765233dc-e1b5-47b8-8eb5-8cef0a62f428","Type":"ContainerStarted","Data":"1fdd6c31d11dd5f6f2d1de0eb9cb22fa38e74c4af1e4da35b278d3c2fc1ba7c5"} Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.447630 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.449205 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"519e0bef7a9547e866c4e60e58aa5143c249078b05d8ca2fcddeeb0b93e508ee"} Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.451648 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" event={"ID":"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2","Type":"ContainerStarted","Data":"4cec178adf4c9faaadaee658d75df0bdeb7062e48673efc1991a7ee3639c642a"} Nov 25 10:02:09 crc kubenswrapper[4926]: I1125 10:02:09.472358 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-bvqpc" podStartSLOduration=1.472331006 podStartE2EDuration="1.472331006s" podCreationTimestamp="2025-11-25 10:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:02:09.468543823 +0000 UTC m=+826.152500270" watchObservedRunningTime="2025-11-25 10:02:09.472331006 +0000 UTC m=+826.156287443" Nov 25 10:02:10 crc kubenswrapper[4926]: I1125 10:02:10.078000 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:10 crc kubenswrapper[4926]: I1125 10:02:10.087618 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7586af8a-a829-4c9f-9589-a3643c147f1d-memberlist\") pod \"speaker-6rxct\" (UID: \"7586af8a-a829-4c9f-9589-a3643c147f1d\") " pod="metallb-system/speaker-6rxct" Nov 25 10:02:10 crc kubenswrapper[4926]: I1125 10:02:10.164635 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-6rxct" Nov 25 10:02:10 crc kubenswrapper[4926]: I1125 10:02:10.464020 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6rxct" event={"ID":"7586af8a-a829-4c9f-9589-a3643c147f1d","Type":"ContainerStarted","Data":"65cda72a5defcddc369d0c262a4de2a2dc97a6b0d569e8b4f869685194aceded"} Nov 25 10:02:10 crc kubenswrapper[4926]: I1125 10:02:10.464561 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6rxct" event={"ID":"7586af8a-a829-4c9f-9589-a3643c147f1d","Type":"ContainerStarted","Data":"1d802acf6b175daf04a5c42eabeba72740ff0e936c31b027a2badf99df16a7f2"} Nov 25 10:02:11 crc kubenswrapper[4926]: I1125 10:02:11.475103 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-6rxct" event={"ID":"7586af8a-a829-4c9f-9589-a3643c147f1d","Type":"ContainerStarted","Data":"a664af7b9841b07767525b5f9b22428730b8e0e12d6c44f7fc1b8c8987b2dfda"} Nov 25 10:02:11 crc kubenswrapper[4926]: I1125 10:02:11.475241 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-6rxct" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.090368 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-6rxct" podStartSLOduration=4.090328894 podStartE2EDuration="4.090328894s" podCreationTimestamp="2025-11-25 10:02:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:02:11.49663628 +0000 UTC m=+828.180592717" watchObservedRunningTime="2025-11-25 10:02:12.090328894 +0000 UTC m=+828.774285321" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.094258 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kg4lh"] Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.095532 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.115540 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kg4lh"] Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.212751 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-catalog-content\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.212836 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-utilities\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.212871 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cv2c\" (UniqueName: \"kubernetes.io/projected/b55e4fdc-411b-4556-a7d2-f340aae947fc-kube-api-access-2cv2c\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.314854 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-catalog-content\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.314980 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-utilities\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.315016 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cv2c\" (UniqueName: \"kubernetes.io/projected/b55e4fdc-411b-4556-a7d2-f340aae947fc-kube-api-access-2cv2c\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.315507 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-utilities\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.315540 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-catalog-content\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.336530 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cv2c\" (UniqueName: \"kubernetes.io/projected/b55e4fdc-411b-4556-a7d2-f340aae947fc-kube-api-access-2cv2c\") pod \"certified-operators-kg4lh\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.416628 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:12 crc kubenswrapper[4926]: I1125 10:02:12.984473 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kg4lh"] Nov 25 10:02:13 crc kubenswrapper[4926]: I1125 10:02:13.500505 4926 generic.go:334] "Generic (PLEG): container finished" podID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerID="f8ad87b065cceb2c3c7b5a09fe465d5cb7039b09c38e60b4d410ad063a81af78" exitCode=0 Nov 25 10:02:13 crc kubenswrapper[4926]: I1125 10:02:13.500586 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerDied","Data":"f8ad87b065cceb2c3c7b5a09fe465d5cb7039b09c38e60b4d410ad063a81af78"} Nov 25 10:02:13 crc kubenswrapper[4926]: I1125 10:02:13.501086 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerStarted","Data":"5a8e7ccd7292640727ef2d4812013e2e4fc14dcd3e6656ce90964359617841d9"} Nov 25 10:02:17 crc kubenswrapper[4926]: I1125 10:02:17.530383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerStarted","Data":"d985a09eb00a0201f48e4f415b98a537277b4071f4c5ecb8d60dd10eda31ff02"} Nov 25 10:02:17 crc kubenswrapper[4926]: I1125 10:02:17.532540 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" event={"ID":"1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2","Type":"ContainerStarted","Data":"1835e723bcb074aa8603ae698fde3122ff20ec3bcb20a6eb0744f794bb29b5fc"} Nov 25 10:02:17 crc kubenswrapper[4926]: I1125 10:02:17.533048 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:17 crc kubenswrapper[4926]: I1125 10:02:17.534239 4926 generic.go:334] "Generic (PLEG): container finished" podID="99b2fe2e-14bb-4dda-9152-bb593b6cb80e" containerID="94e25fdc6a5b6023cc8c817910f0e89542186314bcec4480795ca3811a956051" exitCode=0 Nov 25 10:02:17 crc kubenswrapper[4926]: I1125 10:02:17.534275 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerDied","Data":"94e25fdc6a5b6023cc8c817910f0e89542186314bcec4480795ca3811a956051"} Nov 25 10:02:17 crc kubenswrapper[4926]: I1125 10:02:17.576990 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" podStartSLOduration=1.38505719 podStartE2EDuration="9.576959205s" podCreationTimestamp="2025-11-25 10:02:08 +0000 UTC" firstStartedPulling="2025-11-25 10:02:09.02986312 +0000 UTC m=+825.713819547" lastFinishedPulling="2025-11-25 10:02:17.221765135 +0000 UTC m=+833.905721562" observedRunningTime="2025-11-25 10:02:17.573597503 +0000 UTC m=+834.257553930" watchObservedRunningTime="2025-11-25 10:02:17.576959205 +0000 UTC m=+834.260915632" Nov 25 10:02:18 crc kubenswrapper[4926]: I1125 10:02:18.543311 4926 generic.go:334] "Generic (PLEG): container finished" podID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerID="d985a09eb00a0201f48e4f415b98a537277b4071f4c5ecb8d60dd10eda31ff02" exitCode=0 Nov 25 10:02:18 crc kubenswrapper[4926]: I1125 10:02:18.543416 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerDied","Data":"d985a09eb00a0201f48e4f415b98a537277b4071f4c5ecb8d60dd10eda31ff02"} Nov 25 10:02:18 crc kubenswrapper[4926]: I1125 10:02:18.546289 4926 generic.go:334] "Generic (PLEG): container finished" podID="99b2fe2e-14bb-4dda-9152-bb593b6cb80e" containerID="e2242eba927b299388dc972d535e26c1143ce660cf8e54f6c08b8e8cbe118099" exitCode=0 Nov 25 10:02:18 crc kubenswrapper[4926]: I1125 10:02:18.546375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerDied","Data":"e2242eba927b299388dc972d535e26c1143ce660cf8e54f6c08b8e8cbe118099"} Nov 25 10:02:19 crc kubenswrapper[4926]: I1125 10:02:19.555543 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerStarted","Data":"3ef7a742d2fea28c417f5be0333399895e9fcff1ea7161ae512d595eeb3a4a98"} Nov 25 10:02:19 crc kubenswrapper[4926]: I1125 10:02:19.557812 4926 generic.go:334] "Generic (PLEG): container finished" podID="99b2fe2e-14bb-4dda-9152-bb593b6cb80e" containerID="62fd9b648bbe7c45342b639408b4e51640ee50615d6ae6673b0c3ddd476755c8" exitCode=0 Nov 25 10:02:19 crc kubenswrapper[4926]: I1125 10:02:19.557988 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerDied","Data":"62fd9b648bbe7c45342b639408b4e51640ee50615d6ae6673b0c3ddd476755c8"} Nov 25 10:02:19 crc kubenswrapper[4926]: I1125 10:02:19.586128 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kg4lh" podStartSLOduration=2.146106779 podStartE2EDuration="7.586106868s" podCreationTimestamp="2025-11-25 10:02:12 +0000 UTC" firstStartedPulling="2025-11-25 10:02:13.504131633 +0000 UTC m=+830.188088060" lastFinishedPulling="2025-11-25 10:02:18.944131722 +0000 UTC m=+835.628088149" observedRunningTime="2025-11-25 10:02:19.57842629 +0000 UTC m=+836.262382717" watchObservedRunningTime="2025-11-25 10:02:19.586106868 +0000 UTC m=+836.270063285" Nov 25 10:02:20 crc kubenswrapper[4926]: I1125 10:02:20.169108 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-6rxct" Nov 25 10:02:20 crc kubenswrapper[4926]: I1125 10:02:20.604931 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"62207b617df3164bc367dc62e0884cd28d225770ace8c32d7867b3ec77c3e074"} Nov 25 10:02:20 crc kubenswrapper[4926]: I1125 10:02:20.605587 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"644c9db24083a7fdc61e9a04f021a6c35d166342727ef0f7456ecb4bbf8e85e5"} Nov 25 10:02:20 crc kubenswrapper[4926]: I1125 10:02:20.605610 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"855b715ec18a5d9c6e3f1d14ad16605367526956b215db5808b9407e8f68b396"} Nov 25 10:02:20 crc kubenswrapper[4926]: I1125 10:02:20.605625 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"34f3f6e97e05dd447fb28fe284b48d17c3c8885d5f0635fcd1c1ba8cc29b1b0d"} Nov 25 10:02:20 crc kubenswrapper[4926]: I1125 10:02:20.605639 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"dca25a02c9e433f93c8a33f38bc6eb67a75abd5022b8ab20206af03f1e79cf76"} Nov 25 10:02:21 crc kubenswrapper[4926]: I1125 10:02:21.618316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-49pz2" event={"ID":"99b2fe2e-14bb-4dda-9152-bb593b6cb80e","Type":"ContainerStarted","Data":"c1653f6c600aef4c6fd08d0b9a98117b539698046104648c570850b4c46f9def"} Nov 25 10:02:21 crc kubenswrapper[4926]: I1125 10:02:21.618589 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:21 crc kubenswrapper[4926]: I1125 10:02:21.644655 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-49pz2" podStartSLOduration=5.180884354 podStartE2EDuration="13.644631349s" podCreationTimestamp="2025-11-25 10:02:08 +0000 UTC" firstStartedPulling="2025-11-25 10:02:08.741629779 +0000 UTC m=+825.425586206" lastFinishedPulling="2025-11-25 10:02:17.205376764 +0000 UTC m=+833.889333201" observedRunningTime="2025-11-25 10:02:21.641431211 +0000 UTC m=+838.325387658" watchObservedRunningTime="2025-11-25 10:02:21.644631349 +0000 UTC m=+838.328587776" Nov 25 10:02:22 crc kubenswrapper[4926]: I1125 10:02:22.417435 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:22 crc kubenswrapper[4926]: I1125 10:02:22.417530 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:22 crc kubenswrapper[4926]: I1125 10:02:22.467996 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.108585 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-6ngp4"] Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.109816 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.113049 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.113376 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.113435 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vwm7h" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.141369 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6ngp4"] Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.200746 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2bs6\" (UniqueName: \"kubernetes.io/projected/3cd00b31-399b-4d1e-86d3-1d5f636f5970-kube-api-access-p2bs6\") pod \"openstack-operator-index-6ngp4\" (UID: \"3cd00b31-399b-4d1e-86d3-1d5f636f5970\") " pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.302022 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2bs6\" (UniqueName: \"kubernetes.io/projected/3cd00b31-399b-4d1e-86d3-1d5f636f5970-kube-api-access-p2bs6\") pod \"openstack-operator-index-6ngp4\" (UID: \"3cd00b31-399b-4d1e-86d3-1d5f636f5970\") " pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.328284 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2bs6\" (UniqueName: \"kubernetes.io/projected/3cd00b31-399b-4d1e-86d3-1d5f636f5970-kube-api-access-p2bs6\") pod \"openstack-operator-index-6ngp4\" (UID: \"3cd00b31-399b-4d1e-86d3-1d5f636f5970\") " pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.429914 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.571195 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.615327 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:23 crc kubenswrapper[4926]: I1125 10:02:23.657900 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-6ngp4"] Nov 25 10:02:23 crc kubenswrapper[4926]: W1125 10:02:23.676480 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3cd00b31_399b_4d1e_86d3_1d5f636f5970.slice/crio-6df4caa002240d4d03b4a1eaa81507b6b63bf541146840c629e464b817570911 WatchSource:0}: Error finding container 6df4caa002240d4d03b4a1eaa81507b6b63bf541146840c629e464b817570911: Status 404 returned error can't find the container with id 6df4caa002240d4d03b4a1eaa81507b6b63bf541146840c629e464b817570911 Nov 25 10:02:24 crc kubenswrapper[4926]: I1125 10:02:24.647882 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ngp4" event={"ID":"3cd00b31-399b-4d1e-86d3-1d5f636f5970","Type":"ContainerStarted","Data":"6df4caa002240d4d03b4a1eaa81507b6b63bf541146840c629e464b817570911"} Nov 25 10:02:26 crc kubenswrapper[4926]: I1125 10:02:26.501875 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6ngp4"] Nov 25 10:02:26 crc kubenswrapper[4926]: I1125 10:02:26.661654 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ngp4" event={"ID":"3cd00b31-399b-4d1e-86d3-1d5f636f5970","Type":"ContainerStarted","Data":"a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3"} Nov 25 10:02:26 crc kubenswrapper[4926]: I1125 10:02:26.681420 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-6ngp4" podStartSLOduration=1.18048519 podStartE2EDuration="3.681391994s" podCreationTimestamp="2025-11-25 10:02:23 +0000 UTC" firstStartedPulling="2025-11-25 10:02:23.678768855 +0000 UTC m=+840.362725272" lastFinishedPulling="2025-11-25 10:02:26.179675649 +0000 UTC m=+842.863632076" observedRunningTime="2025-11-25 10:02:26.680567353 +0000 UTC m=+843.364523800" watchObservedRunningTime="2025-11-25 10:02:26.681391994 +0000 UTC m=+843.365348421" Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.112574 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-h4hql"] Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.113694 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.123720 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h4hql"] Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.269672 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8x9z\" (UniqueName: \"kubernetes.io/projected/047b7e2a-f92d-41be-aa69-f3807ae9b0c4-kube-api-access-j8x9z\") pod \"openstack-operator-index-h4hql\" (UID: \"047b7e2a-f92d-41be-aa69-f3807ae9b0c4\") " pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.371549 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8x9z\" (UniqueName: \"kubernetes.io/projected/047b7e2a-f92d-41be-aa69-f3807ae9b0c4-kube-api-access-j8x9z\") pod \"openstack-operator-index-h4hql\" (UID: \"047b7e2a-f92d-41be-aa69-f3807ae9b0c4\") " pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.404529 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8x9z\" (UniqueName: \"kubernetes.io/projected/047b7e2a-f92d-41be-aa69-f3807ae9b0c4-kube-api-access-j8x9z\") pod \"openstack-operator-index-h4hql\" (UID: \"047b7e2a-f92d-41be-aa69-f3807ae9b0c4\") " pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.437613 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.670111 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-6ngp4" podUID="3cd00b31-399b-4d1e-86d3-1d5f636f5970" containerName="registry-server" containerID="cri-o://a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3" gracePeriod=2 Nov 25 10:02:27 crc kubenswrapper[4926]: I1125 10:02:27.853064 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-h4hql"] Nov 25 10:02:27 crc kubenswrapper[4926]: W1125 10:02:27.859389 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod047b7e2a_f92d_41be_aa69_f3807ae9b0c4.slice/crio-bf7e6443dbe50333b16b43af67f919245e145c36199fbac9e554d0dce00e41dd WatchSource:0}: Error finding container bf7e6443dbe50333b16b43af67f919245e145c36199fbac9e554d0dce00e41dd: Status 404 returned error can't find the container with id bf7e6443dbe50333b16b43af67f919245e145c36199fbac9e554d0dce00e41dd Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.592836 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-6ztkb" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.614881 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.679396 4926 generic.go:334] "Generic (PLEG): container finished" podID="3cd00b31-399b-4d1e-86d3-1d5f636f5970" containerID="a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3" exitCode=0 Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.679552 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-6ngp4" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.679540 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ngp4" event={"ID":"3cd00b31-399b-4d1e-86d3-1d5f636f5970","Type":"ContainerDied","Data":"a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3"} Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.679598 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-6ngp4" event={"ID":"3cd00b31-399b-4d1e-86d3-1d5f636f5970","Type":"ContainerDied","Data":"6df4caa002240d4d03b4a1eaa81507b6b63bf541146840c629e464b817570911"} Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.679624 4926 scope.go:117] "RemoveContainer" containerID="a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.684969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h4hql" event={"ID":"047b7e2a-f92d-41be-aa69-f3807ae9b0c4","Type":"ContainerStarted","Data":"bcd95d1badc4e16f780ba4f942ab88592520d73a03fca4dad76297f168549a33"} Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.685024 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-h4hql" event={"ID":"047b7e2a-f92d-41be-aa69-f3807ae9b0c4","Type":"ContainerStarted","Data":"bf7e6443dbe50333b16b43af67f919245e145c36199fbac9e554d0dce00e41dd"} Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.694622 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-bvqpc" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.697838 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2bs6\" (UniqueName: \"kubernetes.io/projected/3cd00b31-399b-4d1e-86d3-1d5f636f5970-kube-api-access-p2bs6\") pod \"3cd00b31-399b-4d1e-86d3-1d5f636f5970\" (UID: \"3cd00b31-399b-4d1e-86d3-1d5f636f5970\") " Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.705522 4926 scope.go:117] "RemoveContainer" containerID="a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3" Nov 25 10:02:28 crc kubenswrapper[4926]: E1125 10:02:28.707007 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3\": container with ID starting with a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3 not found: ID does not exist" containerID="a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.707188 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3"} err="failed to get container status \"a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3\": rpc error: code = NotFound desc = could not find container \"a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3\": container with ID starting with a7eba09a14fff51c0d186461c0a1d8e35a497b7882a728609d597f2421f86fd3 not found: ID does not exist" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.707440 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-h4hql" podStartSLOduration=1.5814468590000001 podStartE2EDuration="1.707426671s" podCreationTimestamp="2025-11-25 10:02:27 +0000 UTC" firstStartedPulling="2025-11-25 10:02:27.863869593 +0000 UTC m=+844.547826020" lastFinishedPulling="2025-11-25 10:02:27.989849405 +0000 UTC m=+844.673805832" observedRunningTime="2025-11-25 10:02:28.699434956 +0000 UTC m=+845.383391403" watchObservedRunningTime="2025-11-25 10:02:28.707426671 +0000 UTC m=+845.391383098" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.709018 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd00b31-399b-4d1e-86d3-1d5f636f5970-kube-api-access-p2bs6" (OuterVolumeSpecName: "kube-api-access-p2bs6") pod "3cd00b31-399b-4d1e-86d3-1d5f636f5970" (UID: "3cd00b31-399b-4d1e-86d3-1d5f636f5970"). InnerVolumeSpecName "kube-api-access-p2bs6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:02:28 crc kubenswrapper[4926]: I1125 10:02:28.799780 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2bs6\" (UniqueName: \"kubernetes.io/projected/3cd00b31-399b-4d1e-86d3-1d5f636f5970-kube-api-access-p2bs6\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:29 crc kubenswrapper[4926]: I1125 10:02:29.023456 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-6ngp4"] Nov 25 10:02:29 crc kubenswrapper[4926]: I1125 10:02:29.027755 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-6ngp4"] Nov 25 10:02:30 crc kubenswrapper[4926]: I1125 10:02:30.020697 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd00b31-399b-4d1e-86d3-1d5f636f5970" path="/var/lib/kubelet/pods/3cd00b31-399b-4d1e-86d3-1d5f636f5970/volumes" Nov 25 10:02:32 crc kubenswrapper[4926]: I1125 10:02:32.475310 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:34 crc kubenswrapper[4926]: I1125 10:02:34.703220 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kg4lh"] Nov 25 10:02:34 crc kubenswrapper[4926]: I1125 10:02:34.705006 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kg4lh" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="registry-server" containerID="cri-o://3ef7a742d2fea28c417f5be0333399895e9fcff1ea7161ae512d595eeb3a4a98" gracePeriod=2 Nov 25 10:02:35 crc kubenswrapper[4926]: I1125 10:02:35.743513 4926 generic.go:334] "Generic (PLEG): container finished" podID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerID="3ef7a742d2fea28c417f5be0333399895e9fcff1ea7161ae512d595eeb3a4a98" exitCode=0 Nov 25 10:02:35 crc kubenswrapper[4926]: I1125 10:02:35.743579 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerDied","Data":"3ef7a742d2fea28c417f5be0333399895e9fcff1ea7161ae512d595eeb3a4a98"} Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.034718 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.112276 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cv2c\" (UniqueName: \"kubernetes.io/projected/b55e4fdc-411b-4556-a7d2-f340aae947fc-kube-api-access-2cv2c\") pod \"b55e4fdc-411b-4556-a7d2-f340aae947fc\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.112375 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-utilities\") pod \"b55e4fdc-411b-4556-a7d2-f340aae947fc\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.112566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-catalog-content\") pod \"b55e4fdc-411b-4556-a7d2-f340aae947fc\" (UID: \"b55e4fdc-411b-4556-a7d2-f340aae947fc\") " Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.113946 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-utilities" (OuterVolumeSpecName: "utilities") pod "b55e4fdc-411b-4556-a7d2-f340aae947fc" (UID: "b55e4fdc-411b-4556-a7d2-f340aae947fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.114229 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.118356 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b55e4fdc-411b-4556-a7d2-f340aae947fc-kube-api-access-2cv2c" (OuterVolumeSpecName: "kube-api-access-2cv2c") pod "b55e4fdc-411b-4556-a7d2-f340aae947fc" (UID: "b55e4fdc-411b-4556-a7d2-f340aae947fc"). InnerVolumeSpecName "kube-api-access-2cv2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.166156 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b55e4fdc-411b-4556-a7d2-f340aae947fc" (UID: "b55e4fdc-411b-4556-a7d2-f340aae947fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.215849 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b55e4fdc-411b-4556-a7d2-f340aae947fc-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.215881 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cv2c\" (UniqueName: \"kubernetes.io/projected/b55e4fdc-411b-4556-a7d2-f340aae947fc-kube-api-access-2cv2c\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.753771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kg4lh" event={"ID":"b55e4fdc-411b-4556-a7d2-f340aae947fc","Type":"ContainerDied","Data":"5a8e7ccd7292640727ef2d4812013e2e4fc14dcd3e6656ce90964359617841d9"} Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.753899 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kg4lh" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.754330 4926 scope.go:117] "RemoveContainer" containerID="3ef7a742d2fea28c417f5be0333399895e9fcff1ea7161ae512d595eeb3a4a98" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.772761 4926 scope.go:117] "RemoveContainer" containerID="d985a09eb00a0201f48e4f415b98a537277b4071f4c5ecb8d60dd10eda31ff02" Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.790083 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kg4lh"] Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.795519 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kg4lh"] Nov 25 10:02:36 crc kubenswrapper[4926]: I1125 10:02:36.807707 4926 scope.go:117] "RemoveContainer" containerID="f8ad87b065cceb2c3c7b5a09fe465d5cb7039b09c38e60b4d410ad063a81af78" Nov 25 10:02:37 crc kubenswrapper[4926]: I1125 10:02:37.438804 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:37 crc kubenswrapper[4926]: I1125 10:02:37.438876 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:37 crc kubenswrapper[4926]: I1125 10:02:37.479532 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:37 crc kubenswrapper[4926]: I1125 10:02:37.793642 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-h4hql" Nov 25 10:02:38 crc kubenswrapper[4926]: I1125 10:02:38.028347 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" path="/var/lib/kubelet/pods/b55e4fdc-411b-4556-a7d2-f340aae947fc/volumes" Nov 25 10:02:38 crc kubenswrapper[4926]: I1125 10:02:38.574163 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-49pz2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.773719 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2"] Nov 25 10:02:44 crc kubenswrapper[4926]: E1125 10:02:44.774817 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="extract-utilities" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.774840 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="extract-utilities" Nov 25 10:02:44 crc kubenswrapper[4926]: E1125 10:02:44.774856 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="extract-content" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.774866 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="extract-content" Nov 25 10:02:44 crc kubenswrapper[4926]: E1125 10:02:44.774895 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="registry-server" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.774903 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="registry-server" Nov 25 10:02:44 crc kubenswrapper[4926]: E1125 10:02:44.774914 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd00b31-399b-4d1e-86d3-1d5f636f5970" containerName="registry-server" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.774922 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd00b31-399b-4d1e-86d3-1d5f636f5970" containerName="registry-server" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.775109 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd00b31-399b-4d1e-86d3-1d5f636f5970" containerName="registry-server" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.775123 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b55e4fdc-411b-4556-a7d2-f340aae947fc" containerName="registry-server" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.776365 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.779396 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-zswrb" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.790882 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2"] Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.863440 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-util\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.863509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-bundle\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.863675 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd7k8\" (UniqueName: \"kubernetes.io/projected/aa8e7b73-9640-4c71-9615-41c32b22f48c-kube-api-access-wd7k8\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.965388 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-util\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.965461 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-bundle\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.965493 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd7k8\" (UniqueName: \"kubernetes.io/projected/aa8e7b73-9640-4c71-9615-41c32b22f48c-kube-api-access-wd7k8\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.965957 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-util\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.966201 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-bundle\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:44 crc kubenswrapper[4926]: I1125 10:02:44.993284 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd7k8\" (UniqueName: \"kubernetes.io/projected/aa8e7b73-9640-4c71-9615-41c32b22f48c-kube-api-access-wd7k8\") pod \"97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:45 crc kubenswrapper[4926]: I1125 10:02:45.103287 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:45 crc kubenswrapper[4926]: I1125 10:02:45.328146 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2"] Nov 25 10:02:45 crc kubenswrapper[4926]: I1125 10:02:45.833162 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerID="5cb68abddb7159679ec3d58a6909c19c57754480d20989522d1ae70a4fde6866" exitCode=0 Nov 25 10:02:45 crc kubenswrapper[4926]: I1125 10:02:45.833235 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" event={"ID":"aa8e7b73-9640-4c71-9615-41c32b22f48c","Type":"ContainerDied","Data":"5cb68abddb7159679ec3d58a6909c19c57754480d20989522d1ae70a4fde6866"} Nov 25 10:02:45 crc kubenswrapper[4926]: I1125 10:02:45.833626 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" event={"ID":"aa8e7b73-9640-4c71-9615-41c32b22f48c","Type":"ContainerStarted","Data":"a81c9574bdaa3816d4156762b79f924015d94f5e603c951efd68f8e18da1bdab"} Nov 25 10:02:47 crc kubenswrapper[4926]: I1125 10:02:47.848970 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerID="907376194da852973a13f0aa43f3ef7ae878f870965623452714a7dcf0743ba1" exitCode=0 Nov 25 10:02:47 crc kubenswrapper[4926]: I1125 10:02:47.849093 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" event={"ID":"aa8e7b73-9640-4c71-9615-41c32b22f48c","Type":"ContainerDied","Data":"907376194da852973a13f0aa43f3ef7ae878f870965623452714a7dcf0743ba1"} Nov 25 10:02:48 crc kubenswrapper[4926]: I1125 10:02:48.861191 4926 generic.go:334] "Generic (PLEG): container finished" podID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerID="a2aaede53d19b7e7962d7f7ca35f450cfc47c15d909c873df72b379e544403c3" exitCode=0 Nov 25 10:02:48 crc kubenswrapper[4926]: I1125 10:02:48.861260 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" event={"ID":"aa8e7b73-9640-4c71-9615-41c32b22f48c","Type":"ContainerDied","Data":"a2aaede53d19b7e7962d7f7ca35f450cfc47c15d909c873df72b379e544403c3"} Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.237984 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.358089 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wd7k8\" (UniqueName: \"kubernetes.io/projected/aa8e7b73-9640-4c71-9615-41c32b22f48c-kube-api-access-wd7k8\") pod \"aa8e7b73-9640-4c71-9615-41c32b22f48c\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.358189 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-bundle\") pod \"aa8e7b73-9640-4c71-9615-41c32b22f48c\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.358228 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-util\") pod \"aa8e7b73-9640-4c71-9615-41c32b22f48c\" (UID: \"aa8e7b73-9640-4c71-9615-41c32b22f48c\") " Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.360618 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-bundle" (OuterVolumeSpecName: "bundle") pod "aa8e7b73-9640-4c71-9615-41c32b22f48c" (UID: "aa8e7b73-9640-4c71-9615-41c32b22f48c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.365988 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa8e7b73-9640-4c71-9615-41c32b22f48c-kube-api-access-wd7k8" (OuterVolumeSpecName: "kube-api-access-wd7k8") pod "aa8e7b73-9640-4c71-9615-41c32b22f48c" (UID: "aa8e7b73-9640-4c71-9615-41c32b22f48c"). InnerVolumeSpecName "kube-api-access-wd7k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.374549 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-util" (OuterVolumeSpecName: "util") pod "aa8e7b73-9640-4c71-9615-41c32b22f48c" (UID: "aa8e7b73-9640-4c71-9615-41c32b22f48c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.459640 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wd7k8\" (UniqueName: \"kubernetes.io/projected/aa8e7b73-9640-4c71-9615-41c32b22f48c-kube-api-access-wd7k8\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.459690 4926 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.459700 4926 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/aa8e7b73-9640-4c71-9615-41c32b22f48c-util\") on node \"crc\" DevicePath \"\"" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.886233 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" event={"ID":"aa8e7b73-9640-4c71-9615-41c32b22f48c","Type":"ContainerDied","Data":"a81c9574bdaa3816d4156762b79f924015d94f5e603c951efd68f8e18da1bdab"} Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.886306 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a81c9574bdaa3816d4156762b79f924015d94f5e603c951efd68f8e18da1bdab" Nov 25 10:02:50 crc kubenswrapper[4926]: I1125 10:02:50.886341 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.083045 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g"] Nov 25 10:02:57 crc kubenswrapper[4926]: E1125 10:02:57.083827 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="util" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.083845 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="util" Nov 25 10:02:57 crc kubenswrapper[4926]: E1125 10:02:57.083861 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="pull" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.083867 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="pull" Nov 25 10:02:57 crc kubenswrapper[4926]: E1125 10:02:57.083897 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="extract" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.083904 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="extract" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.084027 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa8e7b73-9640-4c71-9615-41c32b22f48c" containerName="extract" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.084524 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.088704 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-2fvj4" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.112201 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g"] Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.268607 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl2jj\" (UniqueName: \"kubernetes.io/projected/5cf84563-5163-4108-92dc-1dbf757212de-kube-api-access-vl2jj\") pod \"openstack-operator-controller-operator-6444d758f4-lgw8g\" (UID: \"5cf84563-5163-4108-92dc-1dbf757212de\") " pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.370212 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl2jj\" (UniqueName: \"kubernetes.io/projected/5cf84563-5163-4108-92dc-1dbf757212de-kube-api-access-vl2jj\") pod \"openstack-operator-controller-operator-6444d758f4-lgw8g\" (UID: \"5cf84563-5163-4108-92dc-1dbf757212de\") " pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.393347 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl2jj\" (UniqueName: \"kubernetes.io/projected/5cf84563-5163-4108-92dc-1dbf757212de-kube-api-access-vl2jj\") pod \"openstack-operator-controller-operator-6444d758f4-lgw8g\" (UID: \"5cf84563-5163-4108-92dc-1dbf757212de\") " pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.404572 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.696821 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g"] Nov 25 10:02:57 crc kubenswrapper[4926]: I1125 10:02:57.942110 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" event={"ID":"5cf84563-5163-4108-92dc-1dbf757212de","Type":"ContainerStarted","Data":"93854335d65f6e5dbbcc379bb95531f1f36be41e8ed5a0a86563b33333fc0111"} Nov 25 10:03:01 crc kubenswrapper[4926]: I1125 10:03:01.976285 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" event={"ID":"5cf84563-5163-4108-92dc-1dbf757212de","Type":"ContainerStarted","Data":"720748cdb5ddb350e01205c49d4237fe193664c59c467ad1a9e97aa1241d74df"} Nov 25 10:03:01 crc kubenswrapper[4926]: I1125 10:03:01.977354 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:03:02 crc kubenswrapper[4926]: I1125 10:03:02.012179 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" podStartSLOduration=1.021393716 podStartE2EDuration="5.012145579s" podCreationTimestamp="2025-11-25 10:02:57 +0000 UTC" firstStartedPulling="2025-11-25 10:02:57.720925393 +0000 UTC m=+874.404881830" lastFinishedPulling="2025-11-25 10:03:01.711677266 +0000 UTC m=+878.395633693" observedRunningTime="2025-11-25 10:03:02.007700948 +0000 UTC m=+878.691657385" watchObservedRunningTime="2025-11-25 10:03:02.012145579 +0000 UTC m=+878.696102006" Nov 25 10:03:07 crc kubenswrapper[4926]: I1125 10:03:07.408331 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:03:14 crc kubenswrapper[4926]: I1125 10:03:14.016008 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:03:14 crc kubenswrapper[4926]: I1125 10:03:14.016574 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.055741 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.057629 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.064074 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-llpch" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.065587 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.070586 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.071854 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.076046 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dkvwv" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.081115 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.082257 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.085231 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2m5cf" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.101997 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.116654 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.117907 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.120399 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gb2hj" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.126806 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.139749 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.152457 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dfgr\" (UniqueName: \"kubernetes.io/projected/194893a3-1c67-4a28-a67e-7d7eeef304a7-kube-api-access-4dfgr\") pod \"barbican-operator-controller-manager-86dc4d89c8-fr2mr\" (UID: \"194893a3-1c67-4a28-a67e-7d7eeef304a7\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.167378 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.168815 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.173979 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-lv9rl" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.196638 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.207908 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.208931 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.212111 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-h85ln" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.219696 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.220861 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.225272 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.225382 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-2q4hl" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.225568 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.247194 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.256877 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dfgr\" (UniqueName: \"kubernetes.io/projected/194893a3-1c67-4a28-a67e-7d7eeef304a7-kube-api-access-4dfgr\") pod \"barbican-operator-controller-manager-86dc4d89c8-fr2mr\" (UID: \"194893a3-1c67-4a28-a67e-7d7eeef304a7\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.256931 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpj6s\" (UniqueName: \"kubernetes.io/projected/5249ee3d-0121-42bf-bc4a-7d05a8410a9f-kube-api-access-lpj6s\") pod \"glance-operator-controller-manager-75fd7f55fb-p7v2x\" (UID: \"5249ee3d-0121-42bf-bc4a-7d05a8410a9f\") " pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.257064 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tthb5\" (UniqueName: \"kubernetes.io/projected/09b3e2a6-3249-4ca3-afc9-39dd744ddd10-kube-api-access-tthb5\") pod \"cinder-operator-controller-manager-79856dc55c-67llk\" (UID: \"09b3e2a6-3249-4ca3-afc9-39dd744ddd10\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.257127 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tn2ql\" (UniqueName: \"kubernetes.io/projected/8292ff2d-08ba-4be8-a268-5c8b6d639087-kube-api-access-tn2ql\") pod \"designate-operator-controller-manager-7d695c9b56-4wxmj\" (UID: \"8292ff2d-08ba-4be8-a268-5c8b6d639087\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.260890 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.262752 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.263153 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.276311 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2c78d" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.289312 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.290428 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.304132 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.306251 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rrrds" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.311097 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dfgr\" (UniqueName: \"kubernetes.io/projected/194893a3-1c67-4a28-a67e-7d7eeef304a7-kube-api-access-4dfgr\") pod \"barbican-operator-controller-manager-86dc4d89c8-fr2mr\" (UID: \"194893a3-1c67-4a28-a67e-7d7eeef304a7\") " pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.321505 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.323076 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.326425 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-97ktp" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.335073 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358683 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tthb5\" (UniqueName: \"kubernetes.io/projected/09b3e2a6-3249-4ca3-afc9-39dd744ddd10-kube-api-access-tthb5\") pod \"cinder-operator-controller-manager-79856dc55c-67llk\" (UID: \"09b3e2a6-3249-4ca3-afc9-39dd744ddd10\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358735 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwxmh\" (UniqueName: \"kubernetes.io/projected/1154878f-e8a5-42de-916f-73276b979e74-kube-api-access-mwxmh\") pod \"horizon-operator-controller-manager-68c9694994-kqv6n\" (UID: \"1154878f-e8a5-42de-916f-73276b979e74\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tn2ql\" (UniqueName: \"kubernetes.io/projected/8292ff2d-08ba-4be8-a268-5c8b6d639087-kube-api-access-tn2ql\") pod \"designate-operator-controller-manager-7d695c9b56-4wxmj\" (UID: \"8292ff2d-08ba-4be8-a268-5c8b6d639087\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358806 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67wx2\" (UniqueName: \"kubernetes.io/projected/b6772bac-4625-41d7-915d-2fdabd9916b7-kube-api-access-67wx2\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358860 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8g2m\" (UniqueName: \"kubernetes.io/projected/0ed1bf71-b826-4854-873d-7c6fcf992bbd-kube-api-access-r8g2m\") pod \"ironic-operator-controller-manager-5bfcdc958c-wt256\" (UID: \"0ed1bf71-b826-4854-873d-7c6fcf992bbd\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358882 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpj6s\" (UniqueName: \"kubernetes.io/projected/5249ee3d-0121-42bf-bc4a-7d05a8410a9f-kube-api-access-lpj6s\") pod \"glance-operator-controller-manager-75fd7f55fb-p7v2x\" (UID: \"5249ee3d-0121-42bf-bc4a-7d05a8410a9f\") " pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358898 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.358924 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pc96\" (UniqueName: \"kubernetes.io/projected/df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a-kube-api-access-7pc96\") pod \"heat-operator-controller-manager-774b86978c-4d4g5\" (UID: \"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.378465 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.380244 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.381544 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.383638 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tthb5\" (UniqueName: \"kubernetes.io/projected/09b3e2a6-3249-4ca3-afc9-39dd744ddd10-kube-api-access-tthb5\") pod \"cinder-operator-controller-manager-79856dc55c-67llk\" (UID: \"09b3e2a6-3249-4ca3-afc9-39dd744ddd10\") " pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.384473 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpj6s\" (UniqueName: \"kubernetes.io/projected/5249ee3d-0121-42bf-bc4a-7d05a8410a9f-kube-api-access-lpj6s\") pod \"glance-operator-controller-manager-75fd7f55fb-p7v2x\" (UID: \"5249ee3d-0121-42bf-bc4a-7d05a8410a9f\") " pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.388069 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-k8kl7" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.408205 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.409631 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.410687 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tn2ql\" (UniqueName: \"kubernetes.io/projected/8292ff2d-08ba-4be8-a268-5c8b6d639087-kube-api-access-tn2ql\") pod \"designate-operator-controller-manager-7d695c9b56-4wxmj\" (UID: \"8292ff2d-08ba-4be8-a268-5c8b6d639087\") " pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.411598 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qvsws" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.412696 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.414322 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.430951 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.446194 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.474762 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwxmh\" (UniqueName: \"kubernetes.io/projected/1154878f-e8a5-42de-916f-73276b979e74-kube-api-access-mwxmh\") pod \"horizon-operator-controller-manager-68c9694994-kqv6n\" (UID: \"1154878f-e8a5-42de-916f-73276b979e74\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.474842 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqz8n\" (UniqueName: \"kubernetes.io/projected/5bf64213-939f-4f49-9030-9f9dcc78f8bb-kube-api-access-rqz8n\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-hgcnt\" (UID: \"5bf64213-939f-4f49-9030-9f9dcc78f8bb\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.474906 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67wx2\" (UniqueName: \"kubernetes.io/projected/b6772bac-4625-41d7-915d-2fdabd9916b7-kube-api-access-67wx2\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.474965 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dg69j\" (UniqueName: \"kubernetes.io/projected/dd863871-17e8-4a97-bc8b-5dd02f807225-kube-api-access-dg69j\") pod \"neutron-operator-controller-manager-7c57c8bbc4-947wd\" (UID: \"dd863871-17e8-4a97-bc8b-5dd02f807225\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.475043 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8g2m\" (UniqueName: \"kubernetes.io/projected/0ed1bf71-b826-4854-873d-7c6fcf992bbd-kube-api-access-r8g2m\") pod \"ironic-operator-controller-manager-5bfcdc958c-wt256\" (UID: \"0ed1bf71-b826-4854-873d-7c6fcf992bbd\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.475076 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.475115 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4fn8\" (UniqueName: \"kubernetes.io/projected/bc152fdc-1bab-4528-992d-0883671d7372-kube-api-access-j4fn8\") pod \"keystone-operator-controller-manager-748dc6576f-82vvs\" (UID: \"bc152fdc-1bab-4528-992d-0883671d7372\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.475159 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pc96\" (UniqueName: \"kubernetes.io/projected/df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a-kube-api-access-7pc96\") pod \"heat-operator-controller-manager-774b86978c-4d4g5\" (UID: \"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.475192 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldbvg\" (UniqueName: \"kubernetes.io/projected/409d9035-812c-4b37-b495-d6554315fb03-kube-api-access-ldbvg\") pod \"manila-operator-controller-manager-58bb8d67cc-2b444\" (UID: \"409d9035-812c-4b37-b495-d6554315fb03\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:03:43 crc kubenswrapper[4926]: E1125 10:03:43.476195 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 10:03:43 crc kubenswrapper[4926]: E1125 10:03:43.476264 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert podName:b6772bac-4625-41d7-915d-2fdabd9916b7 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:43.976238731 +0000 UTC m=+920.660195158 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert") pod "infra-operator-controller-manager-858778c9dc-lwgz6" (UID: "b6772bac-4625-41d7-915d-2fdabd9916b7") : secret "infra-operator-webhook-server-cert" not found Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.511893 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67wx2\" (UniqueName: \"kubernetes.io/projected/b6772bac-4625-41d7-915d-2fdabd9916b7-kube-api-access-67wx2\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.514118 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8g2m\" (UniqueName: \"kubernetes.io/projected/0ed1bf71-b826-4854-873d-7c6fcf992bbd-kube-api-access-r8g2m\") pod \"ironic-operator-controller-manager-5bfcdc958c-wt256\" (UID: \"0ed1bf71-b826-4854-873d-7c6fcf992bbd\") " pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.514216 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.515690 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwxmh\" (UniqueName: \"kubernetes.io/projected/1154878f-e8a5-42de-916f-73276b979e74-kube-api-access-mwxmh\") pod \"horizon-operator-controller-manager-68c9694994-kqv6n\" (UID: \"1154878f-e8a5-42de-916f-73276b979e74\") " pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.517403 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pc96\" (UniqueName: \"kubernetes.io/projected/df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a-kube-api-access-7pc96\") pod \"heat-operator-controller-manager-774b86978c-4d4g5\" (UID: \"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a\") " pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.572609 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.576549 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-txn9r" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.580588 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.582067 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4fn8\" (UniqueName: \"kubernetes.io/projected/bc152fdc-1bab-4528-992d-0883671d7372-kube-api-access-j4fn8\") pod \"keystone-operator-controller-manager-748dc6576f-82vvs\" (UID: \"bc152fdc-1bab-4528-992d-0883671d7372\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.582111 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldbvg\" (UniqueName: \"kubernetes.io/projected/409d9035-812c-4b37-b495-d6554315fb03-kube-api-access-ldbvg\") pod \"manila-operator-controller-manager-58bb8d67cc-2b444\" (UID: \"409d9035-812c-4b37-b495-d6554315fb03\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.582139 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqz8n\" (UniqueName: \"kubernetes.io/projected/5bf64213-939f-4f49-9030-9f9dcc78f8bb-kube-api-access-rqz8n\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-hgcnt\" (UID: \"5bf64213-939f-4f49-9030-9f9dcc78f8bb\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.582162 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5nc7m\" (UniqueName: \"kubernetes.io/projected/fa24e0a8-f9a4-4160-bac5-4d73e8579fd8-kube-api-access-5nc7m\") pod \"nova-operator-controller-manager-79556f57fc-tmxx7\" (UID: \"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.582200 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dg69j\" (UniqueName: \"kubernetes.io/projected/dd863871-17e8-4a97-bc8b-5dd02f807225-kube-api-access-dg69j\") pod \"neutron-operator-controller-manager-7c57c8bbc4-947wd\" (UID: \"dd863871-17e8-4a97-bc8b-5dd02f807225\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.594240 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.603319 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.625865 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.635109 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqz8n\" (UniqueName: \"kubernetes.io/projected/5bf64213-939f-4f49-9030-9f9dcc78f8bb-kube-api-access-rqz8n\") pod \"mariadb-operator-controller-manager-cb6c4fdb7-hgcnt\" (UID: \"5bf64213-939f-4f49-9030-9f9dcc78f8bb\") " pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.635539 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.636267 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4fn8\" (UniqueName: \"kubernetes.io/projected/bc152fdc-1bab-4528-992d-0883671d7372-kube-api-access-j4fn8\") pod \"keystone-operator-controller-manager-748dc6576f-82vvs\" (UID: \"bc152fdc-1bab-4528-992d-0883671d7372\") " pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.637023 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dg69j\" (UniqueName: \"kubernetes.io/projected/dd863871-17e8-4a97-bc8b-5dd02f807225-kube-api-access-dg69j\") pod \"neutron-operator-controller-manager-7c57c8bbc4-947wd\" (UID: \"dd863871-17e8-4a97-bc8b-5dd02f807225\") " pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.656220 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rvl6x" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.672768 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldbvg\" (UniqueName: \"kubernetes.io/projected/409d9035-812c-4b37-b495-d6554315fb03-kube-api-access-ldbvg\") pod \"manila-operator-controller-manager-58bb8d67cc-2b444\" (UID: \"409d9035-812c-4b37-b495-d6554315fb03\") " pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.675985 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.679648 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.683765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5nc7m\" (UniqueName: \"kubernetes.io/projected/fa24e0a8-f9a4-4160-bac5-4d73e8579fd8-kube-api-access-5nc7m\") pod \"nova-operator-controller-manager-79556f57fc-tmxx7\" (UID: \"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.723714 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.724816 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.734603 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vvbg6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.756367 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.762079 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.762732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5nc7m\" (UniqueName: \"kubernetes.io/projected/fa24e0a8-f9a4-4160-bac5-4d73e8579fd8-kube-api-access-5nc7m\") pod \"nova-operator-controller-manager-79556f57fc-tmxx7\" (UID: \"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8\") " pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.763229 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.770872 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8r2v8" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.787261 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.788142 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qzkfw\" (UniqueName: \"kubernetes.io/projected/f74db279-9026-4869-a147-d45bf581b618-kube-api-access-qzkfw\") pod \"octavia-operator-controller-manager-fd75fd47d-mt4w9\" (UID: \"f74db279-9026-4869-a147-d45bf581b618\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.789808 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.806048 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.819882 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.822995 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.824199 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.825239 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.825728 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.826814 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.830897 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-r8hb9" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.835980 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-cpxvl" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.845449 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.852259 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.853495 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.861164 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.867751 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.882232 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-mgcqh" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.883022 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.884160 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.886590 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-hrtpv" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.890353 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcc9d\" (UniqueName: \"kubernetes.io/projected/edb138df-3c28-4b7d-b285-2fe43094ead3-kube-api-access-hcc9d\") pod \"ovn-operator-controller-manager-66cf5c67ff-hr2gx\" (UID: \"edb138df-3c28-4b7d-b285-2fe43094ead3\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.890419 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db149e24-86d6-4f34-abb0-82229e25e9d7-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.890447 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwhjf\" (UniqueName: \"kubernetes.io/projected/db149e24-86d6-4f34-abb0-82229e25e9d7-kube-api-access-hwhjf\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.890496 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qzkfw\" (UniqueName: \"kubernetes.io/projected/f74db279-9026-4869-a147-d45bf581b618-kube-api-access-qzkfw\") pod \"octavia-operator-controller-manager-fd75fd47d-mt4w9\" (UID: \"f74db279-9026-4869-a147-d45bf581b618\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.914835 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.928201 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.929390 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qzkfw\" (UniqueName: \"kubernetes.io/projected/f74db279-9026-4869-a147-d45bf581b618-kube-api-access-qzkfw\") pod \"octavia-operator-controller-manager-fd75fd47d-mt4w9\" (UID: \"f74db279-9026-4869-a147-d45bf581b618\") " pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.934278 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.953185 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-vtrb2"] Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.962238 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.968406 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-bhpdd" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.971428 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.992691 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db149e24-86d6-4f34-abb0-82229e25e9d7-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993017 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwhjf\" (UniqueName: \"kubernetes.io/projected/db149e24-86d6-4f34-abb0-82229e25e9d7-kube-api-access-hwhjf\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993054 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5ckj\" (UniqueName: \"kubernetes.io/projected/ae1df4bc-d341-49fd-a49a-e02b4d8714d1-kube-api-access-b5ckj\") pod \"placement-operator-controller-manager-5db546f9d9-lwls8\" (UID: \"ae1df4bc-d341-49fd-a49a-e02b4d8714d1\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993087 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn4t7\" (UniqueName: \"kubernetes.io/projected/416161f0-e9fb-4909-83c0-dfe7ad15885f-kube-api-access-nn4t7\") pod \"swift-operator-controller-manager-6fdc4fcf86-kspmf\" (UID: \"416161f0-e9fb-4909-83c0-dfe7ad15885f\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993124 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9ddb\" (UniqueName: \"kubernetes.io/projected/2494b104-d923-4de0-82e0-8f4ff0e9c5f8-kube-api-access-s9ddb\") pod \"test-operator-controller-manager-5cb74df96-fbbkn\" (UID: \"2494b104-d923-4de0-82e0-8f4ff0e9c5f8\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993159 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993180 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcc9d\" (UniqueName: \"kubernetes.io/projected/edb138df-3c28-4b7d-b285-2fe43094ead3-kube-api-access-hcc9d\") pod \"ovn-operator-controller-manager-66cf5c67ff-hr2gx\" (UID: \"edb138df-3c28-4b7d-b285-2fe43094ead3\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:03:43 crc kubenswrapper[4926]: I1125 10:03:43.993204 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b7nd\" (UniqueName: \"kubernetes.io/projected/7b32eddd-fe58-4854-bc45-a97ec5349f86-kube-api-access-2b7nd\") pod \"telemetry-operator-controller-manager-567f98c9d-5wdlm\" (UID: \"7b32eddd-fe58-4854-bc45-a97ec5349f86\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:03:43 crc kubenswrapper[4926]: E1125 10:03:43.993385 4926 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:03:43 crc kubenswrapper[4926]: E1125 10:03:43.993453 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db149e24-86d6-4f34-abb0-82229e25e9d7-cert podName:db149e24-86d6-4f34-abb0-82229e25e9d7 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:44.493436249 +0000 UTC m=+921.177392676 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/db149e24-86d6-4f34-abb0-82229e25e9d7-cert") pod "openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" (UID: "db149e24-86d6-4f34-abb0-82229e25e9d7") : secret "openstack-baremetal-operator-webhook-server-cert" not found Nov 25 10:03:43 crc kubenswrapper[4926]: E1125 10:03:43.994071 4926 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Nov 25 10:03:43 crc kubenswrapper[4926]: E1125 10:03:43.994105 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert podName:b6772bac-4625-41d7-915d-2fdabd9916b7 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:44.994096075 +0000 UTC m=+921.678052502 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert") pod "infra-operator-controller-manager-858778c9dc-lwgz6" (UID: "b6772bac-4625-41d7-915d-2fdabd9916b7") : secret "infra-operator-webhook-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.002524 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-vtrb2"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.013712 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.015722 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.015780 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.025456 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcc9d\" (UniqueName: \"kubernetes.io/projected/edb138df-3c28-4b7d-b285-2fe43094ead3-kube-api-access-hcc9d\") pod \"ovn-operator-controller-manager-66cf5c67ff-hr2gx\" (UID: \"edb138df-3c28-4b7d-b285-2fe43094ead3\") " pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.027820 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwhjf\" (UniqueName: \"kubernetes.io/projected/db149e24-86d6-4f34-abb0-82229e25e9d7-kube-api-access-hwhjf\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.110116 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.110552 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b7nd\" (UniqueName: \"kubernetes.io/projected/7b32eddd-fe58-4854-bc45-a97ec5349f86-kube-api-access-2b7nd\") pod \"telemetry-operator-controller-manager-567f98c9d-5wdlm\" (UID: \"7b32eddd-fe58-4854-bc45-a97ec5349f86\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.110686 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5ckj\" (UniqueName: \"kubernetes.io/projected/ae1df4bc-d341-49fd-a49a-e02b4d8714d1-kube-api-access-b5ckj\") pod \"placement-operator-controller-manager-5db546f9d9-lwls8\" (UID: \"ae1df4bc-d341-49fd-a49a-e02b4d8714d1\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.110736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn4t7\" (UniqueName: \"kubernetes.io/projected/416161f0-e9fb-4909-83c0-dfe7ad15885f-kube-api-access-nn4t7\") pod \"swift-operator-controller-manager-6fdc4fcf86-kspmf\" (UID: \"416161f0-e9fb-4909-83c0-dfe7ad15885f\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.110835 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9ddb\" (UniqueName: \"kubernetes.io/projected/2494b104-d923-4de0-82e0-8f4ff0e9c5f8-kube-api-access-s9ddb\") pod \"test-operator-controller-manager-5cb74df96-fbbkn\" (UID: \"2494b104-d923-4de0-82e0-8f4ff0e9c5f8\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.110863 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vkxjq\" (UniqueName: \"kubernetes.io/projected/d95f6ac6-9ecf-4cef-ad4e-1389a10f303e-kube-api-access-vkxjq\") pod \"watcher-operator-controller-manager-864885998-vtrb2\" (UID: \"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.114597 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.121948 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.128652 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.129002 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gsprt" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.133076 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.169740 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.188421 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn4t7\" (UniqueName: \"kubernetes.io/projected/416161f0-e9fb-4909-83c0-dfe7ad15885f-kube-api-access-nn4t7\") pod \"swift-operator-controller-manager-6fdc4fcf86-kspmf\" (UID: \"416161f0-e9fb-4909-83c0-dfe7ad15885f\") " pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.189090 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b7nd\" (UniqueName: \"kubernetes.io/projected/7b32eddd-fe58-4854-bc45-a97ec5349f86-kube-api-access-2b7nd\") pod \"telemetry-operator-controller-manager-567f98c9d-5wdlm\" (UID: \"7b32eddd-fe58-4854-bc45-a97ec5349f86\") " pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.189440 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9ddb\" (UniqueName: \"kubernetes.io/projected/2494b104-d923-4de0-82e0-8f4ff0e9c5f8-kube-api-access-s9ddb\") pod \"test-operator-controller-manager-5cb74df96-fbbkn\" (UID: \"2494b104-d923-4de0-82e0-8f4ff0e9c5f8\") " pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.199366 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5ckj\" (UniqueName: \"kubernetes.io/projected/ae1df4bc-d341-49fd-a49a-e02b4d8714d1-kube-api-access-b5ckj\") pod \"placement-operator-controller-manager-5db546f9d9-lwls8\" (UID: \"ae1df4bc-d341-49fd-a49a-e02b4d8714d1\") " pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.205759 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.222789 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.257563 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6kzx\" (UniqueName: \"kubernetes.io/projected/4403c995-1767-4968-b0f4-0a4f0d70a188-kube-api-access-j6kzx\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.257708 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.257773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.257892 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vkxjq\" (UniqueName: \"kubernetes.io/projected/d95f6ac6-9ecf-4cef-ad4e-1389a10f303e-kube-api-access-vkxjq\") pod \"watcher-operator-controller-manager-864885998-vtrb2\" (UID: \"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.261304 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.261491 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.262768 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.271093 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-krm2h" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.279554 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.284114 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vkxjq\" (UniqueName: \"kubernetes.io/projected/d95f6ac6-9ecf-4cef-ad4e-1389a10f303e-kube-api-access-vkxjq\") pod \"watcher-operator-controller-manager-864885998-vtrb2\" (UID: \"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e\") " pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.326328 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.363019 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.363117 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.363222 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6kzx\" (UniqueName: \"kubernetes.io/projected/4403c995-1767-4968-b0f4-0a4f0d70a188-kube-api-access-j6kzx\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.363266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwmwh\" (UniqueName: \"kubernetes.io/projected/03cc50ce-1146-4953-901e-19d5bd7c468e-kube-api-access-lwmwh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-zdqj6\" (UID: \"03cc50ce-1146-4953-901e-19d5bd7c468e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.363474 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.363545 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs podName:4403c995-1767-4968-b0f4-0a4f0d70a188 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:44.863522394 +0000 UTC m=+921.547478821 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs") pod "openstack-operator-controller-manager-86896d4c4b-ssjxq" (UID: "4403c995-1767-4968-b0f4-0a4f0d70a188") : secret "webhook-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.364494 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.364660 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs podName:4403c995-1767-4968-b0f4-0a4f0d70a188 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:44.864591261 +0000 UTC m=+921.548547758 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs") pod "openstack-operator-controller-manager-86896d4c4b-ssjxq" (UID: "4403c995-1767-4968-b0f4-0a4f0d70a188") : secret "metrics-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.385247 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.405019 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6kzx\" (UniqueName: \"kubernetes.io/projected/4403c995-1767-4968-b0f4-0a4f0d70a188-kube-api-access-j6kzx\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.464547 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwmwh\" (UniqueName: \"kubernetes.io/projected/03cc50ce-1146-4953-901e-19d5bd7c468e-kube-api-access-lwmwh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-zdqj6\" (UID: \"03cc50ce-1146-4953-901e-19d5bd7c468e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.491533 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwmwh\" (UniqueName: \"kubernetes.io/projected/03cc50ce-1146-4953-901e-19d5bd7c468e-kube-api-access-lwmwh\") pod \"rabbitmq-cluster-operator-manager-668c99d594-zdqj6\" (UID: \"03cc50ce-1146-4953-901e-19d5bd7c468e\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.520401 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.569773 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db149e24-86d6-4f34-abb0-82229e25e9d7-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.575441 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/db149e24-86d6-4f34-abb0-82229e25e9d7-cert\") pod \"openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs\" (UID: \"db149e24-86d6-4f34-abb0-82229e25e9d7\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.710731 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.724344 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.742505 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x"] Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.764331 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.876579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: I1125 10:03:44.876962 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.877148 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.877233 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs podName:4403c995-1767-4968-b0f4-0a4f0d70a188 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:45.877202175 +0000 UTC m=+922.561158602 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs") pod "openstack-operator-controller-manager-86896d4c4b-ssjxq" (UID: "4403c995-1767-4968-b0f4-0a4f0d70a188") : secret "metrics-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.878029 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:03:44 crc kubenswrapper[4926]: E1125 10:03:44.878074 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs podName:4403c995-1767-4968-b0f4-0a4f0d70a188 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:45.878062047 +0000 UTC m=+922.562018474 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs") pod "openstack-operator-controller-manager-86896d4c4b-ssjxq" (UID: "4403c995-1767-4968-b0f4-0a4f0d70a188") : secret "webhook-server-cert" not found Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.081181 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.084863 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b6772bac-4625-41d7-915d-2fdabd9916b7-cert\") pod \"infra-operator-controller-manager-858778c9dc-lwgz6\" (UID: \"b6772bac-4625-41d7-915d-2fdabd9916b7\") " pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.195216 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.219173 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk"] Nov 25 10:03:45 crc kubenswrapper[4926]: W1125 10:03:45.257128 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod09b3e2a6_3249_4ca3_afc9_39dd744ddd10.slice/crio-336185507dfe5b7e1ec026d5c6cf1520266c4ca03854790a49b553027e82a692 WatchSource:0}: Error finding container 336185507dfe5b7e1ec026d5c6cf1520266c4ca03854790a49b553027e82a692: Status 404 returned error can't find the container with id 336185507dfe5b7e1ec026d5c6cf1520266c4ca03854790a49b553027e82a692 Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.299537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerStarted","Data":"336185507dfe5b7e1ec026d5c6cf1520266c4ca03854790a49b553027e82a692"} Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.308298 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" event={"ID":"194893a3-1c67-4a28-a67e-7d7eeef304a7","Type":"ContainerStarted","Data":"e8545b57840487a104e8677e4f71a0787b1dc0fbe3fa39a4e3078c3be8db0059"} Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.317418 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerStarted","Data":"d7b347620e869439dd9539b049c3c2c8550d5193404bdf981f3843e2b39f17fe"} Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.326883 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerStarted","Data":"92e05ed78c5fa144473bbd3f4dbf0ac86a3975aa728247e4d63368d348dc4965"} Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.346021 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerStarted","Data":"b1dda301fbabb1c399ab425973fc4b6d9046c43511bf090e28cae338f2b2149b"} Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.359252 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.538870 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.565711 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.621721 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.652137 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.663602 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.679660 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.684601 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm"] Nov 25 10:03:45 crc kubenswrapper[4926]: W1125 10:03:45.694930 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc152fdc_1bab_4528_992d_0883671d7372.slice/crio-26b7709e76e75bbeece297579600a62e8c2cde2095401241d2e02535bc411b7d WatchSource:0}: Error finding container 26b7709e76e75bbeece297579600a62e8c2cde2095401241d2e02535bc411b7d: Status 404 returned error can't find the container with id 26b7709e76e75bbeece297579600a62e8c2cde2095401241d2e02535bc411b7d Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.793004 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.822744 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.836556 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.858462 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.875350 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.895978 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7"] Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.903756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.903847 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.904021 4926 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.904113 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs podName:4403c995-1767-4968-b0f4-0a4f0d70a188 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:47.904090306 +0000 UTC m=+924.588046733 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs") pod "openstack-operator-controller-manager-86896d4c4b-ssjxq" (UID: "4403c995-1767-4968-b0f4-0a4f0d70a188") : secret "metrics-server-cert" not found Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.904354 4926 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.904448 4926 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs podName:4403c995-1767-4968-b0f4-0a4f0d70a188 nodeName:}" failed. No retries permitted until 2025-11-25 10:03:47.904418924 +0000 UTC m=+924.588375351 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs") pod "openstack-operator-controller-manager-86896d4c4b-ssjxq" (UID: "4403c995-1767-4968-b0f4-0a4f0d70a188") : secret "webhook-server-cert" not found Nov 25 10:03:45 crc kubenswrapper[4926]: W1125 10:03:45.918721 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03cc50ce_1146_4953_901e_19d5bd7c468e.slice/crio-fb797ab111d2fc3611ba7ff54ad7f3df0390615ad2a9c34e8576dcba439652ac WatchSource:0}: Error finding container fb797ab111d2fc3611ba7ff54ad7f3df0390615ad2a9c34e8576dcba439652ac: Status 404 returned error can't find the container with id fb797ab111d2fc3611ba7ff54ad7f3df0390615ad2a9c34e8576dcba439652ac Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.926476 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9ddb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-fbbkn_openstack-operators(2494b104-d923-4de0-82e0-8f4ff0e9c5f8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: W1125 10:03:45.933586 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb149e24_86d6_4f34_abb0_82229e25e9d7.slice/crio-8ec5b423f96a1623e9d0710ad162912bfc105264d3f3b0963b6225de8ef0b430 WatchSource:0}: Error finding container 8ec5b423f96a1623e9d0710ad162912bfc105264d3f3b0963b6225de8ef0b430: Status 404 returned error can't find the container with id 8ec5b423f96a1623e9d0710ad162912bfc105264d3f3b0963b6225de8ef0b430 Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.937794 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5nc7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-tmxx7_openstack-operators(fa24e0a8-f9a4-4160-bac5-4d73e8579fd8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.937897 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9ddb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5cb74df96-fbbkn_openstack-operators(2494b104-d923-4de0-82e0-8f4ff0e9c5f8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.938082 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwmwh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-zdqj6_openstack-operators(03cc50ce-1146-4953-901e-19d5bd7c468e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.938341 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx"] Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.939413 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" podUID="2494b104-d923-4de0-82e0-8f4ff0e9c5f8" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.939508 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podUID="03cc50ce-1146-4953-901e-19d5bd7c468e" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.941075 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5nc7m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-79556f57fc-tmxx7_openstack-operators(fa24e0a8-f9a4-4160-bac5-4d73e8579fd8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.942229 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" podUID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.946572 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-864885998-vtrb2"] Nov 25 10:03:45 crc kubenswrapper[4926]: W1125 10:03:45.947503 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podedb138df_3c28_4b7d_b285_2fe43094ead3.slice/crio-9574104451f69ab0bdd5a7ef6a98ff9fcb124a88fb304cafe52cb52f4365350b WatchSource:0}: Error finding container 9574104451f69ab0bdd5a7ef6a98ff9fcb124a88fb304cafe52cb52f4365350b: Status 404 returned error can't find the container with id 9574104451f69ab0bdd5a7ef6a98ff9fcb124a88fb304cafe52cb52f4365350b Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.952238 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_API_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-api:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CLOUDKITTY_PROC_IMAGE_URL_DEFAULT,Value:quay.rdoproject.org/podified-master-centos10/openstack-cloudkitty-processor:current,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwhjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs_openstack-operators(db149e24-86d6-4f34-abb0-82229e25e9d7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.952524 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs"] Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.956648 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hwhjf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs_openstack-operators(db149e24-86d6-4f34-abb0-82229e25e9d7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.956838 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hcc9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-hr2gx_openstack-operators(edb138df-3c28-4b7d-b285-2fe43094ead3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.958377 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" podUID="db149e24-86d6-4f34-abb0-82229e25e9d7" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.962006 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hcc9d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-66cf5c67ff-hr2gx_openstack-operators(edb138df-3c28-4b7d-b285-2fe43094ead3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.964170 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" podUID="edb138df-3c28-4b7d-b285-2fe43094ead3" Nov 25 10:03:45 crc kubenswrapper[4926]: W1125 10:03:45.973857 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd95f6ac6_9ecf_4cef_ad4e_1389a10f303e.slice/crio-7ea5d5df454bdacc2d1e31ae269cf062190d451c86bc7a371304f3d723db0c65 WatchSource:0}: Error finding container 7ea5d5df454bdacc2d1e31ae269cf062190d451c86bc7a371304f3d723db0c65: Status 404 returned error can't find the container with id 7ea5d5df454bdacc2d1e31ae269cf062190d451c86bc7a371304f3d723db0c65 Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.987446 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vkxjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-vtrb2_openstack-operators(d95f6ac6-9ecf-4cef-ad4e-1389a10f303e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.990435 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vkxjq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-864885998-vtrb2_openstack-operators(d95f6ac6-9ecf-4cef-ad4e-1389a10f303e): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:45 crc kubenswrapper[4926]: E1125 10:03:45.991744 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" podUID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" Nov 25 10:03:45 crc kubenswrapper[4926]: I1125 10:03:45.995664 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6"] Nov 25 10:03:46 crc kubenswrapper[4926]: W1125 10:03:46.003657 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6772bac_4625_41d7_915d_2fdabd9916b7.slice/crio-01d6cb2d5606ba3cc8216c7933213de7eecf8a34910dac2e8ff0e2e5bcb994d3 WatchSource:0}: Error finding container 01d6cb2d5606ba3cc8216c7933213de7eecf8a34910dac2e8ff0e2e5bcb994d3: Status 404 returned error can't find the container with id 01d6cb2d5606ba3cc8216c7933213de7eecf8a34910dac2e8ff0e2e5bcb994d3 Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.011381 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:f0688f6a55b7b548aaafd5c2c4f0749a43e7ea447c62a24e8b35257c5d8ba17f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-67wx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-858778c9dc-lwgz6_openstack-operators(b6772bac-4625-41d7-915d-2fdabd9916b7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.013633 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-67wx2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-858778c9dc-lwgz6_openstack-operators(b6772bac-4625-41d7-915d-2fdabd9916b7): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.014969 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podUID="b6772bac-4625-41d7-915d-2fdabd9916b7" Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.357437 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" event={"ID":"416161f0-e9fb-4909-83c0-dfe7ad15885f","Type":"ContainerStarted","Data":"27e06bc5f45d8c27ded487cb0976eefcaa7ccf1ca294df8d3664377c90f8c455"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.359798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" event={"ID":"db149e24-86d6-4f34-abb0-82229e25e9d7","Type":"ContainerStarted","Data":"8ec5b423f96a1623e9d0710ad162912bfc105264d3f3b0963b6225de8ef0b430"} Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.362144 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" podUID="db149e24-86d6-4f34-abb0-82229e25e9d7" Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.364994 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerStarted","Data":"ccf1c3d3e96f07498dcb02489bdee729cd8ca0c7fc16af1ca094dcac8bc9cd7b"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.366963 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerStarted","Data":"26b7709e76e75bbeece297579600a62e8c2cde2095401241d2e02535bc411b7d"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.369557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" event={"ID":"03cc50ce-1146-4953-901e-19d5bd7c468e","Type":"ContainerStarted","Data":"fb797ab111d2fc3611ba7ff54ad7f3df0390615ad2a9c34e8576dcba439652ac"} Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.371292 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podUID="03cc50ce-1146-4953-901e-19d5bd7c468e" Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.371654 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerStarted","Data":"dc32f7488e8dca98bbd2ebc34cfc30d24141ae8fc9abfb4b86cdbd4343273c75"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.375265 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" event={"ID":"b6772bac-4625-41d7-915d-2fdabd9916b7","Type":"ContainerStarted","Data":"01d6cb2d5606ba3cc8216c7933213de7eecf8a34910dac2e8ff0e2e5bcb994d3"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.378529 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerStarted","Data":"3b69743ab927adaa50f504a609c2b578f4df6b2d327152a4553cdbf169aed0cb"} Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.381330 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:f0688f6a55b7b548aaafd5c2c4f0749a43e7ea447c62a24e8b35257c5d8ba17f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podUID="b6772bac-4625-41d7-915d-2fdabd9916b7" Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.381681 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" event={"ID":"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e","Type":"ContainerStarted","Data":"7ea5d5df454bdacc2d1e31ae269cf062190d451c86bc7a371304f3d723db0c65"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.383067 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" event={"ID":"edb138df-3c28-4b7d-b285-2fe43094ead3","Type":"ContainerStarted","Data":"9574104451f69ab0bdd5a7ef6a98ff9fcb124a88fb304cafe52cb52f4365350b"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.385370 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" event={"ID":"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8","Type":"ContainerStarted","Data":"9a8abae4569ffcadb40ac0d36d49c648579bc7acfcf1ae317f063a4e232f001b"} Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.386822 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" podUID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.387221 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" podUID="edb138df-3c28-4b7d-b285-2fe43094ead3" Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.387322 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" podUID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.388199 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerStarted","Data":"35d9fd7a8f7630658206ba557c13a198b28bbb27a67a489f4bf780ee975d5e80"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.397174 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" event={"ID":"2494b104-d923-4de0-82e0-8f4ff0e9c5f8","Type":"ContainerStarted","Data":"bca903c4a8c29ed73d1fe1e76c0f2c8166773f10d3dd7ea654ce57a8a9a6d696"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.401698 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" event={"ID":"f74db279-9026-4869-a147-d45bf581b618","Type":"ContainerStarted","Data":"abfcb150ce45615532f0450e87f2b4ac458d3cffa61ad9a66c6d7f3219d2d1c9"} Nov 25 10:03:46 crc kubenswrapper[4926]: E1125 10:03:46.418174 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" podUID="2494b104-d923-4de0-82e0-8f4ff0e9c5f8" Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.423852 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerStarted","Data":"7e3294b386a0ebb01dc5c648367064abb10ecb67c46422145933776b8c0e0a6c"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.427199 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" event={"ID":"8292ff2d-08ba-4be8-a268-5c8b6d639087","Type":"ContainerStarted","Data":"26ecb1d4da86bd8ee6ebf3740e9e122511b7f34595e2bd7a553095ffbcc6e5fc"} Nov 25 10:03:46 crc kubenswrapper[4926]: I1125 10:03:46.428531 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerStarted","Data":"f2bd2f66f9ac62940b4ae89364b4304d5642806079b00105808b91871795aff1"} Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.443138 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:5d49d4594c66eda7b151746cc6e1d3c67c0129b4503eeb043a64ae8ec2da6a1b\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" podUID="edb138df-3c28-4b7d-b285-2fe43094ead3" Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.444322 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podUID="03cc50ce-1146-4953-901e-19d5bd7c468e" Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.445639 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:78852f8ba332a5756c1551c126157f735279101a0fc3277ba4aa4db3478789dd\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" podUID="db149e24-86d6-4f34-abb0-82229e25e9d7" Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.445756 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4838402d41d42c56613d43dc5041aae475a2b18e6172491d6c4d4a78a580697f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" podUID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.445913 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:f0688f6a55b7b548aaafd5c2c4f0749a43e7ea447c62a24e8b35257c5d8ba17f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podUID="b6772bac-4625-41d7-915d-2fdabd9916b7" Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.446634 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:82207e753574d4be246f86c4b074500d66cf20214aa80f0a8525cf3287a35e6d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" podUID="2494b104-d923-4de0-82e0-8f4ff0e9c5f8" Nov 25 10:03:47 crc kubenswrapper[4926]: E1125 10:03:47.446818 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:c053e34316044f14929e16e4f0d97f9f1b24cb68b5e22b925ca74c66aaaed0a7\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" podUID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" Nov 25 10:03:47 crc kubenswrapper[4926]: I1125 10:03:47.941124 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:47 crc kubenswrapper[4926]: I1125 10:03:47.941496 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:47 crc kubenswrapper[4926]: I1125 10:03:47.947452 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-webhook-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:47 crc kubenswrapper[4926]: I1125 10:03:47.947832 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/4403c995-1767-4968-b0f4-0a4f0d70a188-metrics-certs\") pod \"openstack-operator-controller-manager-86896d4c4b-ssjxq\" (UID: \"4403c995-1767-4968-b0f4-0a4f0d70a188\") " pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:48 crc kubenswrapper[4926]: I1125 10:03:48.063744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:54 crc kubenswrapper[4926]: I1125 10:03:54.633240 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq"] Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.000098 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ldbvg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-58bb8d67cc-2b444_openstack-operators(409d9035-812c-4b37-b495-d6554315fb03): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.003311 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.007284 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lpj6s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-75fd7f55fb-p7v2x_openstack-operators(5249ee3d-0121-42bf-bc4a-7d05a8410a9f): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.009278 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.032698 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7pc96,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-774b86978c-4d4g5_openstack-operators(df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.033987 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.517571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" event={"ID":"8292ff2d-08ba-4be8-a268-5c8b6d639087","Type":"ContainerStarted","Data":"49001810fcfbfa6ca05256a32bb07b914fa3e240a471bf9cd81c5af01a0420ba"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.528264 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerStarted","Data":"a255b4fc4d71a17bfa31390573549daada5264701abd9f080f44707ae5b07712"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.537726 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" event={"ID":"f74db279-9026-4869-a147-d45bf581b618","Type":"ContainerStarted","Data":"72594fd7a747457517e5f0c4214007b015ca36bca78226cee674fc1e7d3699f7"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.539156 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerStarted","Data":"a323101e1d43b04f7a693896c68109ac61b3f975edf283514ae070693da13635"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.539920 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.540991 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.548642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerStarted","Data":"8b6b76df23ca2d547afa40eacc6f3bce4d70e0b3f29f406496c7f0cb25b87a79"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.556813 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerStarted","Data":"ef61f457015a05047bb8f43a9915c81c83b01ac25deb26bff52245b232bf70be"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.559930 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerStarted","Data":"d7b64d216e63148d402f7ada0737c596449e9f35674156f9098accb99a55c13f"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.566693 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerStarted","Data":"06dfe4f5a2264dfe2ffe5ca14a1a5014464986fb1fbcd14a804f41cf46f4396e"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.586512 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" event={"ID":"416161f0-e9fb-4909-83c0-dfe7ad15885f","Type":"ContainerStarted","Data":"fd3404684a77a8cab7d504c74b9a9f632755c7c1c63bff3fc75216a3e2b5c822"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.602668 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerStarted","Data":"3b6d22f86522965bb63d95715cfe147c87133d7900a3fc98424c22ab8736cb72"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.613587 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerStarted","Data":"a521e3270b8748427a6452a581a6d324467194c27ccf125143a4ae250a834174"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.616557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerStarted","Data":"b7da2e7b4d7448d9ed118ae4e0a1e12664e5f491ef10105c0c16f7ef07497164"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.616986 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.619410 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.620663 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerStarted","Data":"2d3e6970086339beb02a13daa3467904448359265eaa86ed860edc6c8a7b8ff6"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.624175 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" event={"ID":"4403c995-1767-4968-b0f4-0a4f0d70a188","Type":"ContainerStarted","Data":"ac3f82ab05b703ebd449519e2445a8ba3233383139cdd50b01dec1b6f0cf868b"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.624272 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.624292 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" event={"ID":"4403c995-1767-4968-b0f4-0a4f0d70a188","Type":"ContainerStarted","Data":"9d2a4fbb7201792df7861579b9882b684256106ac5c82b2e636ddedb0314bdc3"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.625897 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" event={"ID":"194893a3-1c67-4a28-a67e-7d7eeef304a7","Type":"ContainerStarted","Data":"cffdc6c8d780ff8826404234f8d89e8422500931f06a3fc4b92d62a5484033b9"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.626871 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerStarted","Data":"9dd1498849e62796c39436f05c633b26a66126e317de44db35d715b1d6a23a22"} Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.627104 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:03:55 crc kubenswrapper[4926]: E1125 10:03:55.631063 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:03:55 crc kubenswrapper[4926]: I1125 10:03:55.675708 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" podStartSLOduration=12.675683238 podStartE2EDuration="12.675683238s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:03:55.671509304 +0000 UTC m=+932.355465731" watchObservedRunningTime="2025-11-25 10:03:55.675683238 +0000 UTC m=+932.359639665" Nov 25 10:03:56 crc kubenswrapper[4926]: E1125 10:03:56.635324 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:03:56 crc kubenswrapper[4926]: E1125 10:03:56.635922 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:03:56 crc kubenswrapper[4926]: E1125 10:03:56.635945 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.675794 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerStarted","Data":"7c386604f5ba10498df1b5685a148f28df9d977971ea021fefebbb4a48e03735"} Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.676577 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.678112 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.680038 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerStarted","Data":"a5d8a4b5e1c1df8c615b23b0e2876b90c0c7cce0bd8d8ce18116d785a3ab1ef6"} Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.680951 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.684080 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerStarted","Data":"d6c790459225bd6991b001a8d7cee749a46de25a498b1580079e2d5bcfb7031d"} Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.684269 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.686025 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" event={"ID":"8292ff2d-08ba-4be8-a268-5c8b6d639087","Type":"ContainerStarted","Data":"29398410d09e57bc881309295c64869dd9a88daa672425c22d50731644389d84"} Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.686480 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.693757 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.693986 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.694967 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.725466 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" podStartSLOduration=3.119505979 podStartE2EDuration="18.725443981s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.682493248 +0000 UTC m=+922.366449675" lastFinishedPulling="2025-11-25 10:04:01.28843125 +0000 UTC m=+937.972387677" observedRunningTime="2025-11-25 10:04:01.69643872 +0000 UTC m=+938.380395177" watchObservedRunningTime="2025-11-25 10:04:01.725443981 +0000 UTC m=+938.409400408" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.748374 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" podStartSLOduration=3.4143082 podStartE2EDuration="18.748355052s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.870318675 +0000 UTC m=+922.554275102" lastFinishedPulling="2025-11-25 10:04:01.204365527 +0000 UTC m=+937.888321954" observedRunningTime="2025-11-25 10:04:01.747676165 +0000 UTC m=+938.431632612" watchObservedRunningTime="2025-11-25 10:04:01.748355052 +0000 UTC m=+938.432311479" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.752978 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" podStartSLOduration=2.251618338 podStartE2EDuration="18.752967107s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:44.833457706 +0000 UTC m=+921.517414133" lastFinishedPulling="2025-11-25 10:04:01.334806485 +0000 UTC m=+938.018762902" observedRunningTime="2025-11-25 10:04:01.724153049 +0000 UTC m=+938.408109476" watchObservedRunningTime="2025-11-25 10:04:01.752967107 +0000 UTC m=+938.436923524" Nov 25 10:04:01 crc kubenswrapper[4926]: I1125 10:04:01.794909 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" podStartSLOduration=3.141985898 podStartE2EDuration="18.79488721s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.574268823 +0000 UTC m=+922.258225250" lastFinishedPulling="2025-11-25 10:04:01.227170135 +0000 UTC m=+937.911126562" observedRunningTime="2025-11-25 10:04:01.794473 +0000 UTC m=+938.478429427" watchObservedRunningTime="2025-11-25 10:04:01.79488721 +0000 UTC m=+938.478843647" Nov 25 10:04:02 crc kubenswrapper[4926]: I1125 10:04:02.700152 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerStarted","Data":"13857980eea3eb431c825a7de566afd4370650233f9cad22ba3b9849e5d3ac61"} Nov 25 10:04:02 crc kubenswrapper[4926]: I1125 10:04:02.721016 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" podStartSLOduration=3.655061624 podStartE2EDuration="19.720990061s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.28059537 +0000 UTC m=+921.964551797" lastFinishedPulling="2025-11-25 10:04:01.346523807 +0000 UTC m=+938.030480234" observedRunningTime="2025-11-25 10:04:02.71854005 +0000 UTC m=+939.402496477" watchObservedRunningTime="2025-11-25 10:04:02.720990061 +0000 UTC m=+939.404946498" Nov 25 10:04:03 crc kubenswrapper[4926]: I1125 10:04:03.413423 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:04:03 crc kubenswrapper[4926]: I1125 10:04:03.416656 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:04:03 crc kubenswrapper[4926]: I1125 10:04:03.453994 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:04:03 crc kubenswrapper[4926]: I1125 10:04:03.795157 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:04:03 crc kubenswrapper[4926]: I1125 10:04:03.974881 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.074960 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.756793 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerStarted","Data":"295eeb53084407aa4240447654c8ca2ec0eca12ae0415570542f7b3d5af25a9b"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.761777 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerStarted","Data":"817e85c14e469ad64efc686e4a0da90d24009d08d5fcac1a24a6124e7c8f405e"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.763669 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.769240 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.779642 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerStarted","Data":"aa7b04776a2c7e80fef85ee9d83948881daa2db7355249a89716606e28d82476"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.780717 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.785135 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.785597 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podStartSLOduration=17.404985521 podStartE2EDuration="25.785585746s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.835045567 +0000 UTC m=+922.519001994" lastFinishedPulling="2025-11-25 10:03:54.215645792 +0000 UTC m=+930.899602219" observedRunningTime="2025-11-25 10:04:08.785455853 +0000 UTC m=+945.469412280" watchObservedRunningTime="2025-11-25 10:04:08.785585746 +0000 UTC m=+945.469542173" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.790811 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerStarted","Data":"b51119a7b7912ce55f10f44fea583f85d262527dd55121006efbc54eca7b99e4"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.792609 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.795267 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.795801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" event={"ID":"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e","Type":"ContainerStarted","Data":"ff03ca3433f656dcc4c1212ad13eb83d6942deb52316809074bd63004d06bd25"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.795827 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" event={"ID":"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e","Type":"ContainerStarted","Data":"9fcc3ce783d8ba3ae2c657c88c82388db47bbbb8acf55e583dcf67a89db85724"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.796243 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.809278 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" event={"ID":"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8","Type":"ContainerStarted","Data":"a4cc372a5a23eaef51017517d72d3e1cb64f16ad97fdea996f35765ac422e162"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.809334 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" event={"ID":"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8","Type":"ContainerStarted","Data":"ed19bb0315e4fa43211796d7d47cb107e576e80156b84d62a677779e9547cb96"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.810110 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.827404 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" podStartSLOduration=3.4691790080000002 podStartE2EDuration="25.827381817s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.23879477 +0000 UTC m=+921.922751197" lastFinishedPulling="2025-11-25 10:04:07.596997579 +0000 UTC m=+944.280954006" observedRunningTime="2025-11-25 10:04:08.826894985 +0000 UTC m=+945.510851412" watchObservedRunningTime="2025-11-25 10:04:08.827381817 +0000 UTC m=+945.511338244" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.834464 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerStarted","Data":"2b5969a764f5ea8d0f68cbb193c791e69e9f7267148a167c8f721b2a8c9f2f5f"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.850339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" event={"ID":"edb138df-3c28-4b7d-b285-2fe43094ead3","Type":"ContainerStarted","Data":"3397d5d8c101283fd9768218dcfb92f22323a35a77ce26283523f9aa28211925"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.850394 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" event={"ID":"edb138df-3c28-4b7d-b285-2fe43094ead3","Type":"ContainerStarted","Data":"1fff1d775d23d3f0cc426529f74c311a05723ee591a8f9e68adf4b7f0899a62f"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.851372 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.859530 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" podStartSLOduration=5.514344625 podStartE2EDuration="25.859507707s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.673484684 +0000 UTC m=+922.357441121" lastFinishedPulling="2025-11-25 10:04:06.018647776 +0000 UTC m=+942.702604203" observedRunningTime="2025-11-25 10:04:08.856289467 +0000 UTC m=+945.540245894" watchObservedRunningTime="2025-11-25 10:04:08.859507707 +0000 UTC m=+945.543464134" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.859850 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" event={"ID":"2494b104-d923-4de0-82e0-8f4ff0e9c5f8","Type":"ContainerStarted","Data":"89643c7b90b220c650837dba970489dbd4fc1313f213b80b4c37c10f1d8f6ee1"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.859903 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" event={"ID":"2494b104-d923-4de0-82e0-8f4ff0e9c5f8","Type":"ContainerStarted","Data":"77d1e8202eba21168861272b6c126891e870b62c52177ff38897d61849fed639"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.860606 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.868940 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" event={"ID":"db149e24-86d6-4f34-abb0-82229e25e9d7","Type":"ContainerStarted","Data":"24e1f4e3c5c9981b5cd50667f043ba7450c3ec7cecebe478f43d1499d59a0b2d"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.869020 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" event={"ID":"db149e24-86d6-4f34-abb0-82229e25e9d7","Type":"ContainerStarted","Data":"5efd45a115f512a93e860eca1c1f1d00774a08e91e9be3c03ebf65fecffd6f90"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.869217 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.889023 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" event={"ID":"b6772bac-4625-41d7-915d-2fdabd9916b7","Type":"ContainerStarted","Data":"4319fe8a09276249c9c0dd5f1374ff69e21a713688c8c3edaece9a964c68d0fb"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.889082 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" event={"ID":"b6772bac-4625-41d7-915d-2fdabd9916b7","Type":"ContainerStarted","Data":"10425a230b9a32fe43c39a81bf425675d2869d9c23771ebaafa529e3812ee55e"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.889876 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.900310 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" podStartSLOduration=8.376870273 podStartE2EDuration="25.900290251s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.956752538 +0000 UTC m=+922.640708965" lastFinishedPulling="2025-11-25 10:04:03.480172516 +0000 UTC m=+940.164128943" observedRunningTime="2025-11-25 10:04:08.897260716 +0000 UTC m=+945.581217153" watchObservedRunningTime="2025-11-25 10:04:08.900290251 +0000 UTC m=+945.584246678" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.900389 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" event={"ID":"f74db279-9026-4869-a147-d45bf581b618","Type":"ContainerStarted","Data":"20723c84854d1adda313d7b06dfaec63f03fffeb8a94c2374711ee16b43b3f1b"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.901268 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.918191 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" event={"ID":"194893a3-1c67-4a28-a67e-7d7eeef304a7","Type":"ContainerStarted","Data":"e12bea2896dc9ba88b08603b35ef655d4750a7f66ebb6bca7f99bceb6ddd6be8"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.918286 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.919129 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.931172 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.937864 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" event={"ID":"416161f0-e9fb-4909-83c0-dfe7ad15885f","Type":"ContainerStarted","Data":"d9283fe8aa1cef5ce96f7564cd6b96bd406950c6805a8353755c75c8597d0c66"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.938912 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.952457 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.961136 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerStarted","Data":"1e1fdd09264f9083740cb763497ffa41d847e97e36b8be8f3e49b3bd1a9bc57a"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.973425 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" podStartSLOduration=4.349345644 podStartE2EDuration="25.973399632s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.987279078 +0000 UTC m=+922.671235505" lastFinishedPulling="2025-11-25 10:04:07.611333066 +0000 UTC m=+944.295289493" observedRunningTime="2025-11-25 10:04:08.939240081 +0000 UTC m=+945.623196528" watchObservedRunningTime="2025-11-25 10:04:08.973399632 +0000 UTC m=+945.657356059" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.978039 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerStarted","Data":"d07072af1582ff0947d57ff841e733d296a00305be16508ad721551d5b3de1b8"} Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.978901 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.981234 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podStartSLOduration=17.478121051 podStartE2EDuration="25.981205446s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.674211202 +0000 UTC m=+922.358167629" lastFinishedPulling="2025-11-25 10:03:54.177295597 +0000 UTC m=+930.861252024" observedRunningTime="2025-11-25 10:04:08.971500095 +0000 UTC m=+945.655456542" watchObservedRunningTime="2025-11-25 10:04:08.981205446 +0000 UTC m=+945.665161873" Nov 25 10:04:08 crc kubenswrapper[4926]: I1125 10:04:08.987356 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.005008 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" event={"ID":"03cc50ce-1146-4953-901e-19d5bd7c468e","Type":"ContainerStarted","Data":"0017977077d412e6e68756b8087f27244ca4dd25c08e12d1c1a6d9f6b09ce1c1"} Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.033532 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" podStartSLOduration=4.156696387 podStartE2EDuration="26.033508459s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.71669933 +0000 UTC m=+922.400655767" lastFinishedPulling="2025-11-25 10:04:07.593511412 +0000 UTC m=+944.277467839" observedRunningTime="2025-11-25 10:04:09.002361253 +0000 UTC m=+945.686317690" watchObservedRunningTime="2025-11-25 10:04:09.033508459 +0000 UTC m=+945.717464886" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.054531 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podStartSLOduration=4.943192161 podStartE2EDuration="26.054512662s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:46.011163072 +0000 UTC m=+922.695119499" lastFinishedPulling="2025-11-25 10:04:07.122483573 +0000 UTC m=+943.806440000" observedRunningTime="2025-11-25 10:04:09.050483931 +0000 UTC m=+945.734440368" watchObservedRunningTime="2025-11-25 10:04:09.054512662 +0000 UTC m=+945.738469089" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.055533 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" podStartSLOduration=4.393402991 podStartE2EDuration="26.055527137s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.937640692 +0000 UTC m=+922.621597119" lastFinishedPulling="2025-11-25 10:04:07.599764828 +0000 UTC m=+944.283721265" observedRunningTime="2025-11-25 10:04:09.030373961 +0000 UTC m=+945.714330388" watchObservedRunningTime="2025-11-25 10:04:09.055527137 +0000 UTC m=+945.739483564" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.098874 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" podStartSLOduration=4.453517868 podStartE2EDuration="26.098850196s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.951535528 +0000 UTC m=+922.635491955" lastFinishedPulling="2025-11-25 10:04:07.596867856 +0000 UTC m=+944.280824283" observedRunningTime="2025-11-25 10:04:09.090455027 +0000 UTC m=+945.774411444" watchObservedRunningTime="2025-11-25 10:04:09.098850196 +0000 UTC m=+945.782806623" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.114925 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podStartSLOduration=3.358033591 podStartE2EDuration="25.114899606s" podCreationTimestamp="2025-11-25 10:03:44 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.93796117 +0000 UTC m=+922.621917597" lastFinishedPulling="2025-11-25 10:04:07.694827185 +0000 UTC m=+944.378783612" observedRunningTime="2025-11-25 10:04:09.111601423 +0000 UTC m=+945.795557850" watchObservedRunningTime="2025-11-25 10:04:09.114899606 +0000 UTC m=+945.798856033" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.167692 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" podStartSLOduration=6.584411019 podStartE2EDuration="26.16767153s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:44.816666878 +0000 UTC m=+921.500623305" lastFinishedPulling="2025-11-25 10:04:04.399927389 +0000 UTC m=+941.083883816" observedRunningTime="2025-11-25 10:04:09.143944129 +0000 UTC m=+945.827900556" watchObservedRunningTime="2025-11-25 10:04:09.16767153 +0000 UTC m=+945.851627957" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.201486 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" podStartSLOduration=4.996853827 podStartE2EDuration="26.201460951s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.917984692 +0000 UTC m=+922.601941119" lastFinishedPulling="2025-11-25 10:04:07.122591786 +0000 UTC m=+943.806548243" observedRunningTime="2025-11-25 10:04:09.20060477 +0000 UTC m=+945.884561197" watchObservedRunningTime="2025-11-25 10:04:09.201460951 +0000 UTC m=+945.885417368" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.203992 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" podStartSLOduration=6.952131405 podStartE2EDuration="26.203983944s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.674273553 +0000 UTC m=+922.358229980" lastFinishedPulling="2025-11-25 10:04:04.926126092 +0000 UTC m=+941.610082519" observedRunningTime="2025-11-25 10:04:09.171379762 +0000 UTC m=+945.855336189" watchObservedRunningTime="2025-11-25 10:04:09.203983944 +0000 UTC m=+945.887940371" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.228310 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" podStartSLOduration=4.332142735 podStartE2EDuration="26.228289099s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.684294603 +0000 UTC m=+922.368251030" lastFinishedPulling="2025-11-25 10:04:07.580440967 +0000 UTC m=+944.264397394" observedRunningTime="2025-11-25 10:04:09.22592294 +0000 UTC m=+945.909879367" watchObservedRunningTime="2025-11-25 10:04:09.228289099 +0000 UTC m=+945.912245526" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.281798 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" podStartSLOduration=6.160871663 podStartE2EDuration="26.281774981s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:45.919333726 +0000 UTC m=+922.603290153" lastFinishedPulling="2025-11-25 10:04:06.040237044 +0000 UTC m=+942.724193471" observedRunningTime="2025-11-25 10:04:09.253415285 +0000 UTC m=+945.937371712" watchObservedRunningTime="2025-11-25 10:04:09.281774981 +0000 UTC m=+945.965731408" Nov 25 10:04:09 crc kubenswrapper[4926]: I1125 10:04:09.283226 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podStartSLOduration=16.872794078 podStartE2EDuration="26.283220237s" podCreationTimestamp="2025-11-25 10:03:43 +0000 UTC" firstStartedPulling="2025-11-25 10:03:44.823658832 +0000 UTC m=+921.507615259" lastFinishedPulling="2025-11-25 10:03:54.234084991 +0000 UTC m=+930.918041418" observedRunningTime="2025-11-25 10:04:09.282803176 +0000 UTC m=+945.966759603" watchObservedRunningTime="2025-11-25 10:04:09.283220237 +0000 UTC m=+945.967176664" Nov 25 10:04:13 crc kubenswrapper[4926]: I1125 10:04:13.938140 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.020578 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.020637 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.028785 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.029842 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bc7b6ba06aa7d0baa2979a51d5458380a1834ddc41f7963299878af0ce453ec"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.030000 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://9bc7b6ba06aa7d0baa2979a51d5458380a1834ddc41f7963299878af0ce453ec" gracePeriod=600 Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.116643 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.328985 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.389849 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:04:14 crc kubenswrapper[4926]: I1125 10:04:14.769905 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:04:15 crc kubenswrapper[4926]: I1125 10:04:15.056409 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="9bc7b6ba06aa7d0baa2979a51d5458380a1834ddc41f7963299878af0ce453ec" exitCode=0 Nov 25 10:04:15 crc kubenswrapper[4926]: I1125 10:04:15.056471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"9bc7b6ba06aa7d0baa2979a51d5458380a1834ddc41f7963299878af0ce453ec"} Nov 25 10:04:15 crc kubenswrapper[4926]: I1125 10:04:15.056516 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"bdee8ff08446a7fcbf9166a91fe5650194045a7c04583b6368b152639d328c44"} Nov 25 10:04:15 crc kubenswrapper[4926]: I1125 10:04:15.056544 4926 scope.go:117] "RemoveContainer" containerID="d32c8b7deb1af6488ea2af61857b46c8de27ff7dc9d3d581bdb4b8b8f28e8724" Nov 25 10:04:15 crc kubenswrapper[4926]: I1125 10:04:15.368626 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.955307 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wtwj2"] Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.956976 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.960868 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-76lks" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.960957 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.961272 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.961527 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.961724 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 10:04:29 crc kubenswrapper[4926]: I1125 10:04:29.966411 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wtwj2"] Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.047134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-config\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.047183 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksx9r\" (UniqueName: \"kubernetes.io/projected/b2a7fb75-7611-42d1-b363-1da8314440a7-kube-api-access-ksx9r\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.047298 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.148798 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.148890 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-config\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.148926 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksx9r\" (UniqueName: \"kubernetes.io/projected/b2a7fb75-7611-42d1-b363-1da8314440a7-kube-api-access-ksx9r\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.149661 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.150297 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-config\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.170869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksx9r\" (UniqueName: \"kubernetes.io/projected/b2a7fb75-7611-42d1-b363-1da8314440a7-kube-api-access-ksx9r\") pod \"dnsmasq-dns-78dd6ddcc-wtwj2\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.279512 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.735805 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wtwj2"] Nov 25 10:04:30 crc kubenswrapper[4926]: I1125 10:04:30.744630 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:04:31 crc kubenswrapper[4926]: I1125 10:04:31.187823 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" event={"ID":"b2a7fb75-7611-42d1-b363-1da8314440a7","Type":"ContainerStarted","Data":"2dc1bd5058981b4ec53bf45d1d21a2111cb5945ec11ce22d3e4335a617aae55a"} Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.087177 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-b4qlp"] Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.089626 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.140579 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-b4qlp"] Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.197614 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vzdm\" (UniqueName: \"kubernetes.io/projected/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-kube-api-access-6vzdm\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.197675 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-config\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.197703 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.299255 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.299405 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vzdm\" (UniqueName: \"kubernetes.io/projected/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-kube-api-access-6vzdm\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.299448 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-config\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.300675 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-config\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.301378 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-dns-svc\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.332520 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vzdm\" (UniqueName: \"kubernetes.io/projected/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-kube-api-access-6vzdm\") pod \"dnsmasq-dns-666b6646f7-b4qlp\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.384698 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wtwj2"] Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.410047 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.415708 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhgqq"] Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.417185 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.434013 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhgqq"] Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.502820 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-config\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.503135 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvc84\" (UniqueName: \"kubernetes.io/projected/43ea3e47-5f76-4668-84d7-dae46d461369-kube-api-access-bvc84\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.503205 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.604699 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.604780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-config\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.604805 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvc84\" (UniqueName: \"kubernetes.io/projected/43ea3e47-5f76-4668-84d7-dae46d461369-kube-api-access-bvc84\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.605724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.605838 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-config\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.647323 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvc84\" (UniqueName: \"kubernetes.io/projected/43ea3e47-5f76-4668-84d7-dae46d461369-kube-api-access-bvc84\") pod \"dnsmasq-dns-57d769cc4f-fhgqq\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:33 crc kubenswrapper[4926]: I1125 10:04:33.741695 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.245591 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.247027 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.249216 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.255095 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.255350 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ddhxn" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.255606 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.255773 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.256012 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.256214 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.262667 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314235 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314361 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64dbdcf7-3a3c-48c3-a68f-22644f250edd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314393 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-config-data\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314419 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314452 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314480 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grgn6\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-kube-api-access-grgn6\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314510 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314532 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64dbdcf7-3a3c-48c3-a68f-22644f250edd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314556 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314577 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.314632 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.415979 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416042 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416113 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64dbdcf7-3a3c-48c3-a68f-22644f250edd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416137 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-config-data\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416160 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416193 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416217 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grgn6\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-kube-api-access-grgn6\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416241 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416261 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416280 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.416299 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64dbdcf7-3a3c-48c3-a68f-22644f250edd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.417898 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.418214 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.418299 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.418456 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.418471 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-config-data\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.419874 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-server-conf\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.424144 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.426059 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64dbdcf7-3a3c-48c3-a68f-22644f250edd-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.426810 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.433907 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64dbdcf7-3a3c-48c3-a68f-22644f250edd-pod-info\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.435909 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grgn6\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-kube-api-access-grgn6\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.438021 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.521993 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.523603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.527821 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.532449 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.533143 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.534142 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4plth" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.535213 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.535554 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.535745 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.541213 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.633229 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.633796 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.633870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.633969 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634004 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634028 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634066 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634089 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qg2n\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-kube-api-access-6qg2n\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634235 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78f39106-b4c8-41d0-b852-a12e64f5863b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634499 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78f39106-b4c8-41d0-b852-a12e64f5863b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.634535 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.735721 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736653 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736683 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qg2n\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-kube-api-access-6qg2n\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736738 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736781 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78f39106-b4c8-41d0-b852-a12e64f5863b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736816 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78f39106-b4c8-41d0-b852-a12e64f5863b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736840 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736874 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736892 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.737537 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.736567 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.737781 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.738767 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.738856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.739856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.746542 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78f39106-b4c8-41d0-b852-a12e64f5863b-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.747036 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78f39106-b4c8-41d0-b852-a12e64f5863b-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.747149 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.748916 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.765105 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qg2n\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-kube-api-access-6qg2n\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.767223 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:34 crc kubenswrapper[4926]: I1125 10:04:34.849680 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.169390 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.172164 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.174685 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-bqb7m" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.177265 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.185919 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.186003 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.193883 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.196915 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.258782 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-kolla-config\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.258848 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b065a-b959-4809-8f8d-1a36b09579c9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.258880 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.259100 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2a7b065a-b959-4809-8f8d-1a36b09579c9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.259144 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.259167 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-config-data-default\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.259194 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a7b065a-b959-4809-8f8d-1a36b09579c9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.259222 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pg8bq\" (UniqueName: \"kubernetes.io/projected/2a7b065a-b959-4809-8f8d-1a36b09579c9-kube-api-access-pg8bq\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.360964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361025 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2a7b065a-b959-4809-8f8d-1a36b09579c9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361070 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-config-data-default\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361120 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a7b065a-b959-4809-8f8d-1a36b09579c9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361152 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pg8bq\" (UniqueName: \"kubernetes.io/projected/2a7b065a-b959-4809-8f8d-1a36b09579c9-kube-api-access-pg8bq\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361190 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-kolla-config\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361220 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b065a-b959-4809-8f8d-1a36b09579c9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361563 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.361650 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2a7b065a-b959-4809-8f8d-1a36b09579c9-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.362334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-config-data-default\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.362879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.362910 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2a7b065a-b959-4809-8f8d-1a36b09579c9-kolla-config\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.372380 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2a7b065a-b959-4809-8f8d-1a36b09579c9-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.372492 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a7b065a-b959-4809-8f8d-1a36b09579c9-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.389577 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pg8bq\" (UniqueName: \"kubernetes.io/projected/2a7b065a-b959-4809-8f8d-1a36b09579c9-kube-api-access-pg8bq\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.395875 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-galera-0\" (UID: \"2a7b065a-b959-4809-8f8d-1a36b09579c9\") " pod="openstack/openstack-galera-0" Nov 25 10:04:36 crc kubenswrapper[4926]: I1125 10:04:36.507318 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.457795 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.460334 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.463604 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-rr5v7" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.469706 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.469892 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.469992 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.476530 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.578799 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zqh8\" (UniqueName: \"kubernetes.io/projected/156bc1bd-066f-4071-8d4a-b2124fe381bb-kube-api-access-9zqh8\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.578851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/156bc1bd-066f-4071-8d4a-b2124fe381bb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.578888 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156bc1bd-066f-4071-8d4a-b2124fe381bb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.578947 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/156bc1bd-066f-4071-8d4a-b2124fe381bb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.579228 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.579384 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.579443 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.579489 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680788 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680854 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zqh8\" (UniqueName: \"kubernetes.io/projected/156bc1bd-066f-4071-8d4a-b2124fe381bb-kube-api-access-9zqh8\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680882 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/156bc1bd-066f-4071-8d4a-b2124fe381bb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680911 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156bc1bd-066f-4071-8d4a-b2124fe381bb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.680963 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/156bc1bd-066f-4071-8d4a-b2124fe381bb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.681015 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.681163 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.682027 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.682047 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.682180 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/156bc1bd-066f-4071-8d4a-b2124fe381bb-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.682809 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/156bc1bd-066f-4071-8d4a-b2124fe381bb-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.687030 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/156bc1bd-066f-4071-8d4a-b2124fe381bb-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.692880 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/156bc1bd-066f-4071-8d4a-b2124fe381bb-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.700334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zqh8\" (UniqueName: \"kubernetes.io/projected/156bc1bd-066f-4071-8d4a-b2124fe381bb-kube-api-access-9zqh8\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.716165 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"openstack-cell1-galera-0\" (UID: \"156bc1bd-066f-4071-8d4a-b2124fe381bb\") " pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.779861 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.780909 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.783203 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.783536 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-6g2nb" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.784045 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.801226 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.803025 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.884639 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd48201c-f958-4647-afe6-055c685cb255-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.884735 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd48201c-f958-4647-afe6-055c685cb255-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.884915 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnhnt\" (UniqueName: \"kubernetes.io/projected/dd48201c-f958-4647-afe6-055c685cb255-kube-api-access-fnhnt\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.885138 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd48201c-f958-4647-afe6-055c685cb255-config-data\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.885190 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd48201c-f958-4647-afe6-055c685cb255-kolla-config\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.987062 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd48201c-f958-4647-afe6-055c685cb255-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.987120 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnhnt\" (UniqueName: \"kubernetes.io/projected/dd48201c-f958-4647-afe6-055c685cb255-kube-api-access-fnhnt\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.987195 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd48201c-f958-4647-afe6-055c685cb255-config-data\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.987224 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd48201c-f958-4647-afe6-055c685cb255-kolla-config\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.987254 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd48201c-f958-4647-afe6-055c685cb255-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.988422 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/dd48201c-f958-4647-afe6-055c685cb255-kolla-config\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.988567 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dd48201c-f958-4647-afe6-055c685cb255-config-data\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.992559 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dd48201c-f958-4647-afe6-055c685cb255-combined-ca-bundle\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:37 crc kubenswrapper[4926]: I1125 10:04:37.993206 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/dd48201c-f958-4647-afe6-055c685cb255-memcached-tls-certs\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:38 crc kubenswrapper[4926]: I1125 10:04:38.015717 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnhnt\" (UniqueName: \"kubernetes.io/projected/dd48201c-f958-4647-afe6-055c685cb255-kube-api-access-fnhnt\") pod \"memcached-0\" (UID: \"dd48201c-f958-4647-afe6-055c685cb255\") " pod="openstack/memcached-0" Nov 25 10:04:38 crc kubenswrapper[4926]: I1125 10:04:38.112850 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.398158 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.399307 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.401632 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cdkvl" Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.411808 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.516169 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fc9k\" (UniqueName: \"kubernetes.io/projected/435aea2d-06c8-43f7-a905-9b1f9e2da38f-kube-api-access-6fc9k\") pod \"kube-state-metrics-0\" (UID: \"435aea2d-06c8-43f7-a905-9b1f9e2da38f\") " pod="openstack/kube-state-metrics-0" Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.617806 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fc9k\" (UniqueName: \"kubernetes.io/projected/435aea2d-06c8-43f7-a905-9b1f9e2da38f-kube-api-access-6fc9k\") pod \"kube-state-metrics-0\" (UID: \"435aea2d-06c8-43f7-a905-9b1f9e2da38f\") " pod="openstack/kube-state-metrics-0" Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.647366 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fc9k\" (UniqueName: \"kubernetes.io/projected/435aea2d-06c8-43f7-a905-9b1f9e2da38f-kube-api-access-6fc9k\") pod \"kube-state-metrics-0\" (UID: \"435aea2d-06c8-43f7-a905-9b1f9e2da38f\") " pod="openstack/kube-state-metrics-0" Nov 25 10:04:39 crc kubenswrapper[4926]: I1125 10:04:39.731249 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.584063 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6dkhn"] Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.585365 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.589561 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bswhl" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.589670 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.589791 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.595096 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6dkhn"] Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.600575 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-dkcjx"] Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.605002 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678627 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-run-ovn\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678686 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-etc-ovs\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678710 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-log\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678736 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-lib\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678765 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-run\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678804 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fll48\" (UniqueName: \"kubernetes.io/projected/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-kube-api-access-fll48\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678846 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-run\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-log-ovn\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678887 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v746w\" (UniqueName: \"kubernetes.io/projected/af231b27-f9dd-466d-83d0-05484183099f-kube-api-access-v746w\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678922 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af231b27-f9dd-466d-83d0-05484183099f-scripts\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678966 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/af231b27-f9dd-466d-83d0-05484183099f-ovn-controller-tls-certs\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.678997 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af231b27-f9dd-466d-83d0-05484183099f-combined-ca-bundle\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.679016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-scripts\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.682255 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dkcjx"] Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780692 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/af231b27-f9dd-466d-83d0-05484183099f-ovn-controller-tls-certs\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780765 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af231b27-f9dd-466d-83d0-05484183099f-combined-ca-bundle\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780798 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-scripts\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780849 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-run-ovn\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780878 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-etc-ovs\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-log\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780931 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-lib\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.780977 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-run\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.781022 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fll48\" (UniqueName: \"kubernetes.io/projected/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-kube-api-access-fll48\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.781065 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-run\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.781091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-log-ovn\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.781116 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v746w\" (UniqueName: \"kubernetes.io/projected/af231b27-f9dd-466d-83d0-05484183099f-kube-api-access-v746w\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.781153 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af231b27-f9dd-466d-83d0-05484183099f-scripts\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.782886 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-log\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.782963 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-lib\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.783027 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-etc-ovs\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.783118 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-run-ovn\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.783160 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-var-run\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.783337 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-run\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.783490 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/af231b27-f9dd-466d-83d0-05484183099f-var-log-ovn\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.784555 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af231b27-f9dd-466d-83d0-05484183099f-scripts\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.785076 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-scripts\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.786758 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af231b27-f9dd-466d-83d0-05484183099f-combined-ca-bundle\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.795575 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/af231b27-f9dd-466d-83d0-05484183099f-ovn-controller-tls-certs\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.797001 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fll48\" (UniqueName: \"kubernetes.io/projected/dbdb45ad-32a5-4edc-aaa3-d907b5272b8a-kube-api-access-fll48\") pod \"ovn-controller-ovs-dkcjx\" (UID: \"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a\") " pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.798901 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v746w\" (UniqueName: \"kubernetes.io/projected/af231b27-f9dd-466d-83d0-05484183099f-kube-api-access-v746w\") pod \"ovn-controller-6dkhn\" (UID: \"af231b27-f9dd-466d-83d0-05484183099f\") " pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.954344 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.969249 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.991466 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.993878 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.997818 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.998138 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9dq2x" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.998482 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.999157 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 10:04:43 crc kubenswrapper[4926]: I1125 10:04:43.999286 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.007215 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.091368 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7753413-a221-4236-9003-e722280e15bf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.091452 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tsm7n\" (UniqueName: \"kubernetes.io/projected/e7753413-a221-4236-9003-e722280e15bf-kube-api-access-tsm7n\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.091628 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7753413-a221-4236-9003-e722280e15bf-config\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.092160 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7753413-a221-4236-9003-e722280e15bf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.092366 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.092602 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.092714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.092975 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.195856 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7753413-a221-4236-9003-e722280e15bf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.195910 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tsm7n\" (UniqueName: \"kubernetes.io/projected/e7753413-a221-4236-9003-e722280e15bf-kube-api-access-tsm7n\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.195967 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7753413-a221-4236-9003-e722280e15bf-config\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.195991 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7753413-a221-4236-9003-e722280e15bf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.196027 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.196054 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.196078 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.196135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.196987 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/e7753413-a221-4236-9003-e722280e15bf-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.197175 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.197521 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7753413-a221-4236-9003-e722280e15bf-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.198110 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7753413-a221-4236-9003-e722280e15bf-config\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.202044 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.202250 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.202691 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/e7753413-a221-4236-9003-e722280e15bf-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.219582 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.226827 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tsm7n\" (UniqueName: \"kubernetes.io/projected/e7753413-a221-4236-9003-e722280e15bf-kube-api-access-tsm7n\") pod \"ovsdbserver-sb-0\" (UID: \"e7753413-a221-4236-9003-e722280e15bf\") " pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:44 crc kubenswrapper[4926]: I1125 10:04:44.321756 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.079718 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.082386 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.085811 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.086053 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.086176 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-zmqcr" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.086362 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.096549 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132205 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt59n\" (UniqueName: \"kubernetes.io/projected/17396f6f-d876-49fe-9294-fe074f9da844-kube-api-access-bt59n\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132244 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132273 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17396f6f-d876-49fe-9294-fe074f9da844-config\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132304 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17396f6f-d876-49fe-9294-fe074f9da844-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132320 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132337 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132412 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.132459 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17396f6f-d876-49fe-9294-fe074f9da844-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234459 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234541 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17396f6f-d876-49fe-9294-fe074f9da844-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234607 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt59n\" (UniqueName: \"kubernetes.io/projected/17396f6f-d876-49fe-9294-fe074f9da844-kube-api-access-bt59n\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234630 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234660 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17396f6f-d876-49fe-9294-fe074f9da844-config\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234688 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17396f6f-d876-49fe-9294-fe074f9da844-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234710 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.234737 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.235130 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.235468 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/17396f6f-d876-49fe-9294-fe074f9da844-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.235733 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/17396f6f-d876-49fe-9294-fe074f9da844-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.236282 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/17396f6f-d876-49fe-9294-fe074f9da844-config\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.243482 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.244244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.254389 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt59n\" (UniqueName: \"kubernetes.io/projected/17396f6f-d876-49fe-9294-fe074f9da844-kube-api-access-bt59n\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.263835 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/17396f6f-d876-49fe-9294-fe074f9da844-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.272303 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ovsdbserver-nb-0\" (UID: \"17396f6f-d876-49fe-9294-fe074f9da844\") " pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.418211 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Nov 25 10:04:46 crc kubenswrapper[4926]: I1125 10:04:46.425659 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-b4qlp"] Nov 25 10:04:47 crc kubenswrapper[4926]: W1125 10:04:47.058488 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4859cc0d_8bb6_40fc_aa73_f7d2567b314f.slice/crio-44b68d190c018715f762faea0ed9bb2b493eec326cb73d70a123a114f46ff512 WatchSource:0}: Error finding container 44b68d190c018715f762faea0ed9bb2b493eec326cb73d70a123a114f46ff512: Status 404 returned error can't find the container with id 44b68d190c018715f762faea0ed9bb2b493eec326cb73d70a123a114f46ff512 Nov 25 10:04:47 crc kubenswrapper[4926]: E1125 10:04:47.072618 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Nov 25 10:04:47 crc kubenswrapper[4926]: E1125 10:04:47.072975 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ksx9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-wtwj2_openstack(b2a7fb75-7611-42d1-b363-1da8314440a7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:04:47 crc kubenswrapper[4926]: E1125 10:04:47.074671 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" podUID="b2a7fb75-7611-42d1-b363-1da8314440a7" Nov 25 10:04:47 crc kubenswrapper[4926]: I1125 10:04:47.357811 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" event={"ID":"4859cc0d-8bb6-40fc-aa73-f7d2567b314f","Type":"ContainerStarted","Data":"44b68d190c018715f762faea0ed9bb2b493eec326cb73d70a123a114f46ff512"} Nov 25 10:04:47 crc kubenswrapper[4926]: I1125 10:04:47.768557 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Nov 25 10:04:47 crc kubenswrapper[4926]: I1125 10:04:47.799013 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:04:47 crc kubenswrapper[4926]: W1125 10:04:47.825175 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64dbdcf7_3a3c_48c3_a68f_22644f250edd.slice/crio-e23c16e3d6fc2c9d531dc78028ed4c4b1663c115d3156f4b0597313ca1684e6e WatchSource:0}: Error finding container e23c16e3d6fc2c9d531dc78028ed4c4b1663c115d3156f4b0597313ca1684e6e: Status 404 returned error can't find the container with id e23c16e3d6fc2c9d531dc78028ed4c4b1663c115d3156f4b0597313ca1684e6e Nov 25 10:04:47 crc kubenswrapper[4926]: I1125 10:04:47.956608 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.078009 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-dns-svc\") pod \"b2a7fb75-7611-42d1-b363-1da8314440a7\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.078163 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-config\") pod \"b2a7fb75-7611-42d1-b363-1da8314440a7\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.078218 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ksx9r\" (UniqueName: \"kubernetes.io/projected/b2a7fb75-7611-42d1-b363-1da8314440a7-kube-api-access-ksx9r\") pod \"b2a7fb75-7611-42d1-b363-1da8314440a7\" (UID: \"b2a7fb75-7611-42d1-b363-1da8314440a7\") " Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.078782 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-config" (OuterVolumeSpecName: "config") pod "b2a7fb75-7611-42d1-b363-1da8314440a7" (UID: "b2a7fb75-7611-42d1-b363-1da8314440a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.078795 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2a7fb75-7611-42d1-b363-1da8314440a7" (UID: "b2a7fb75-7611-42d1-b363-1da8314440a7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.097842 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2a7fb75-7611-42d1-b363-1da8314440a7-kube-api-access-ksx9r" (OuterVolumeSpecName: "kube-api-access-ksx9r") pod "b2a7fb75-7611-42d1-b363-1da8314440a7" (UID: "b2a7fb75-7611-42d1-b363-1da8314440a7"). InnerVolumeSpecName "kube-api-access-ksx9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.134885 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhgqq"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.134930 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6dkhn"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.139441 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.147114 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.155049 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.165273 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.180298 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.180335 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2a7fb75-7611-42d1-b363-1da8314440a7-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.180349 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ksx9r\" (UniqueName: \"kubernetes.io/projected/b2a7fb75-7611-42d1-b363-1da8314440a7-kube-api-access-ksx9r\") on node \"crc\" DevicePath \"\"" Nov 25 10:04:48 crc kubenswrapper[4926]: W1125 10:04:48.251143 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7753413_a221_4236_9003_e722280e15bf.slice/crio-76b8b70237225f686a2738a31aff0e4440413ca03d3aaea678fb3a9528efded7 WatchSource:0}: Error finding container 76b8b70237225f686a2738a31aff0e4440413ca03d3aaea678fb3a9528efded7: Status 404 returned error can't find the container with id 76b8b70237225f686a2738a31aff0e4440413ca03d3aaea678fb3a9528efded7 Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.255544 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.333089 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-dkcjx"] Nov 25 10:04:48 crc kubenswrapper[4926]: W1125 10:04:48.340189 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbdb45ad_32a5_4edc_aaa3_d907b5272b8a.slice/crio-4536baf4f5fb4f1c3a79612cdd113157d196df8b1d5cf99398f4b256fec351d4 WatchSource:0}: Error finding container 4536baf4f5fb4f1c3a79612cdd113157d196df8b1d5cf99398f4b256fec351d4: Status 404 returned error can't find the container with id 4536baf4f5fb4f1c3a79612cdd113157d196df8b1d5cf99398f4b256fec351d4 Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.366264 4926 generic.go:334] "Generic (PLEG): container finished" podID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerID="4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924" exitCode=0 Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.366362 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" event={"ID":"4859cc0d-8bb6-40fc-aa73-f7d2567b314f","Type":"ContainerDied","Data":"4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.367672 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkcjx" event={"ID":"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a","Type":"ContainerStarted","Data":"4536baf4f5fb4f1c3a79612cdd113157d196df8b1d5cf99398f4b256fec351d4"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.368756 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2a7b065a-b959-4809-8f8d-1a36b09579c9","Type":"ContainerStarted","Data":"7273299de5d1eac0fde087153938f028d3e777b5dac900e0b026bcd2d32cce2c"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.370800 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78f39106-b4c8-41d0-b852-a12e64f5863b","Type":"ContainerStarted","Data":"6bd78357f61cc706e1109a39b225e8e3b5c0bf417d721046b11af1a0957312b6"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.375153 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" event={"ID":"43ea3e47-5f76-4668-84d7-dae46d461369","Type":"ContainerStarted","Data":"5a930120db87ed74e121c2fd0e9d988ae28ea4ea5cda0432e0961ac96a22abfa"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.377870 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" event={"ID":"b2a7fb75-7611-42d1-b363-1da8314440a7","Type":"ContainerDied","Data":"2dc1bd5058981b4ec53bf45d1d21a2111cb5945ec11ce22d3e4335a617aae55a"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.378003 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-wtwj2" Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.384153 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"156bc1bd-066f-4071-8d4a-b2124fe381bb","Type":"ContainerStarted","Data":"8d9a8b1aa46dbb3a096126a2d268fd5930b58dc3afc5cadec24769fab1def9e2"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.385410 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e7753413-a221-4236-9003-e722280e15bf","Type":"ContainerStarted","Data":"76b8b70237225f686a2738a31aff0e4440413ca03d3aaea678fb3a9528efded7"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.395298 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"64dbdcf7-3a3c-48c3-a68f-22644f250edd","Type":"ContainerStarted","Data":"e23c16e3d6fc2c9d531dc78028ed4c4b1663c115d3156f4b0597313ca1684e6e"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.397700 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dd48201c-f958-4647-afe6-055c685cb255","Type":"ContainerStarted","Data":"23e4086ea3330621daa2da09b2d9e40fd5594cddb465fda424770a6b63b39dda"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.398730 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6dkhn" event={"ID":"af231b27-f9dd-466d-83d0-05484183099f","Type":"ContainerStarted","Data":"5acd25ef95bbe0b0e61bb86ec38e5733975178a8130cc7223ccc6da4c7c47c2d"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.401301 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435aea2d-06c8-43f7-a905-9b1f9e2da38f","Type":"ContainerStarted","Data":"eac837a6422e6254720852dca60ae55f196ae93fd4eee28cb330469c1da7086a"} Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.439114 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wtwj2"] Nov 25 10:04:48 crc kubenswrapper[4926]: I1125 10:04:48.445259 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-wtwj2"] Nov 25 10:04:48 crc kubenswrapper[4926]: E1125 10:04:48.535503 4926 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Nov 25 10:04:48 crc kubenswrapper[4926]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/4859cc0d-8bb6-40fc-aa73-f7d2567b314f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 10:04:48 crc kubenswrapper[4926]: > podSandboxID="44b68d190c018715f762faea0ed9bb2b493eec326cb73d70a123a114f46ff512" Nov 25 10:04:48 crc kubenswrapper[4926]: E1125 10:04:48.535690 4926 kuberuntime_manager.go:1274] "Unhandled Error" err=< Nov 25 10:04:48 crc kubenswrapper[4926]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6vzdm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-666b6646f7-b4qlp_openstack(4859cc0d-8bb6-40fc-aa73-f7d2567b314f): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/4859cc0d-8bb6-40fc-aa73-f7d2567b314f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Nov 25 10:04:48 crc kubenswrapper[4926]: > logger="UnhandledError" Nov 25 10:04:48 crc kubenswrapper[4926]: E1125 10:04:48.537100 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/4859cc0d-8bb6-40fc-aa73-f7d2567b314f/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" Nov 25 10:04:49 crc kubenswrapper[4926]: I1125 10:04:49.176064 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Nov 25 10:04:49 crc kubenswrapper[4926]: W1125 10:04:49.349827 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17396f6f_d876_49fe_9294_fe074f9da844.slice/crio-0f51027ba6d1480aa951a8525382dafc0d3d5ed39430f32809733e98a42d4291 WatchSource:0}: Error finding container 0f51027ba6d1480aa951a8525382dafc0d3d5ed39430f32809733e98a42d4291: Status 404 returned error can't find the container with id 0f51027ba6d1480aa951a8525382dafc0d3d5ed39430f32809733e98a42d4291 Nov 25 10:04:49 crc kubenswrapper[4926]: I1125 10:04:49.422362 4926 generic.go:334] "Generic (PLEG): container finished" podID="43ea3e47-5f76-4668-84d7-dae46d461369" containerID="32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e" exitCode=0 Nov 25 10:04:49 crc kubenswrapper[4926]: I1125 10:04:49.423453 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" event={"ID":"43ea3e47-5f76-4668-84d7-dae46d461369","Type":"ContainerDied","Data":"32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e"} Nov 25 10:04:49 crc kubenswrapper[4926]: I1125 10:04:49.430194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"17396f6f-d876-49fe-9294-fe074f9da844","Type":"ContainerStarted","Data":"0f51027ba6d1480aa951a8525382dafc0d3d5ed39430f32809733e98a42d4291"} Nov 25 10:04:50 crc kubenswrapper[4926]: I1125 10:04:50.026554 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2a7fb75-7611-42d1-b363-1da8314440a7" path="/var/lib/kubelet/pods/b2a7fb75-7611-42d1-b363-1da8314440a7/volumes" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.501640 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435aea2d-06c8-43f7-a905-9b1f9e2da38f","Type":"ContainerStarted","Data":"255f0b47abd114ce9e1d3b22dcecd051f3fe06d6de05d51e5bedcbc583028585"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.502507 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.505403 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"dd48201c-f958-4647-afe6-055c685cb255","Type":"ContainerStarted","Data":"49be58eb8e0bf02d7db27413750f9010d5cd44809c2b7e2335d009c882f03c12"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.505522 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.507994 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"156bc1bd-066f-4071-8d4a-b2124fe381bb","Type":"ContainerStarted","Data":"e9f840bbf118def8a2a798831c802b438ec563ccdefb3f8865aeead33c20991f"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.510771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2a7b065a-b959-4809-8f8d-1a36b09579c9","Type":"ContainerStarted","Data":"71927ca595a6b92131f2576fb3500de8b9537a676f72172a6f019c95912005e4"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.514884 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6dkhn" event={"ID":"af231b27-f9dd-466d-83d0-05484183099f","Type":"ContainerStarted","Data":"182ebe87dccb300d9199127032d3dbcbb4827dacc02a83e6d9d5367961c7b575"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.515043 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-6dkhn" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.516854 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkcjx" event={"ID":"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a","Type":"ContainerStarted","Data":"38e04f2665b3bca645a89235bd1aa7b8f2178f3dfebaed8bae40ea2dba04806a"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.519402 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"17396f6f-d876-49fe-9294-fe074f9da844","Type":"ContainerStarted","Data":"f0679fa9affb5fdc4c1fe6c676186999eb47c740848204f710e5b65d4fb5e7fa"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.525103 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.101951622 podStartE2EDuration="18.525088036s" podCreationTimestamp="2025-11-25 10:04:39 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.176025427 +0000 UTC m=+984.859981854" lastFinishedPulling="2025-11-25 10:04:56.599161841 +0000 UTC m=+993.283118268" observedRunningTime="2025-11-25 10:04:57.520906452 +0000 UTC m=+994.204862879" watchObservedRunningTime="2025-11-25 10:04:57.525088036 +0000 UTC m=+994.209044463" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.536894 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" event={"ID":"43ea3e47-5f76-4668-84d7-dae46d461369","Type":"ContainerStarted","Data":"7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.537147 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.539610 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e7753413-a221-4236-9003-e722280e15bf","Type":"ContainerStarted","Data":"da9f3ce7f7740df5ec265d8a19903dae247293af7597a4b2728bcb1ba05f2f3f"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.542409 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" event={"ID":"4859cc0d-8bb6-40fc-aa73-f7d2567b314f","Type":"ContainerStarted","Data":"b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7"} Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.542686 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.581005 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=12.706279843 podStartE2EDuration="20.580981393s" podCreationTimestamp="2025-11-25 10:04:37 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.18181585 +0000 UTC m=+984.865772277" lastFinishedPulling="2025-11-25 10:04:56.0565174 +0000 UTC m=+992.740473827" observedRunningTime="2025-11-25 10:04:57.576486832 +0000 UTC m=+994.260443279" watchObservedRunningTime="2025-11-25 10:04:57.580981393 +0000 UTC m=+994.264937820" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.651773 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-6dkhn" podStartSLOduration=6.755420824 podStartE2EDuration="14.65173361s" podCreationTimestamp="2025-11-25 10:04:43 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.160339567 +0000 UTC m=+984.844295994" lastFinishedPulling="2025-11-25 10:04:56.056652353 +0000 UTC m=+992.740608780" observedRunningTime="2025-11-25 10:04:57.646426108 +0000 UTC m=+994.330382545" watchObservedRunningTime="2025-11-25 10:04:57.65173361 +0000 UTC m=+994.335690037" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.665173 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" podStartSLOduration=24.210020825 podStartE2EDuration="24.665154973s" podCreationTimestamp="2025-11-25 10:04:33 +0000 UTC" firstStartedPulling="2025-11-25 10:04:47.077269131 +0000 UTC m=+983.761225578" lastFinishedPulling="2025-11-25 10:04:47.532403309 +0000 UTC m=+984.216359726" observedRunningTime="2025-11-25 10:04:57.663461911 +0000 UTC m=+994.347418338" watchObservedRunningTime="2025-11-25 10:04:57.665154973 +0000 UTC m=+994.349111400" Nov 25 10:04:57 crc kubenswrapper[4926]: I1125 10:04:57.696224 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" podStartSLOduration=24.696202264 podStartE2EDuration="24.696202264s" podCreationTimestamp="2025-11-25 10:04:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:04:57.682923204 +0000 UTC m=+994.366879631" watchObservedRunningTime="2025-11-25 10:04:57.696202264 +0000 UTC m=+994.380158691" Nov 25 10:04:58 crc kubenswrapper[4926]: I1125 10:04:58.552440 4926 generic.go:334] "Generic (PLEG): container finished" podID="dbdb45ad-32a5-4edc-aaa3-d907b5272b8a" containerID="38e04f2665b3bca645a89235bd1aa7b8f2178f3dfebaed8bae40ea2dba04806a" exitCode=0 Nov 25 10:04:58 crc kubenswrapper[4926]: I1125 10:04:58.552773 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkcjx" event={"ID":"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a","Type":"ContainerDied","Data":"38e04f2665b3bca645a89235bd1aa7b8f2178f3dfebaed8bae40ea2dba04806a"} Nov 25 10:04:58 crc kubenswrapper[4926]: I1125 10:04:58.557254 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78f39106-b4c8-41d0-b852-a12e64f5863b","Type":"ContainerStarted","Data":"838134005ec8486202b3ca3e5e28bed22a4b79db2906d9166c0e208a2fcef3c5"} Nov 25 10:04:58 crc kubenswrapper[4926]: I1125 10:04:58.558827 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"64dbdcf7-3a3c-48c3-a68f-22644f250edd","Type":"ContainerStarted","Data":"07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b"} Nov 25 10:05:01 crc kubenswrapper[4926]: I1125 10:05:01.590525 4926 generic.go:334] "Generic (PLEG): container finished" podID="2a7b065a-b959-4809-8f8d-1a36b09579c9" containerID="71927ca595a6b92131f2576fb3500de8b9537a676f72172a6f019c95912005e4" exitCode=0 Nov 25 10:05:01 crc kubenswrapper[4926]: I1125 10:05:01.590588 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2a7b065a-b959-4809-8f8d-1a36b09579c9","Type":"ContainerDied","Data":"71927ca595a6b92131f2576fb3500de8b9537a676f72172a6f019c95912005e4"} Nov 25 10:05:01 crc kubenswrapper[4926]: I1125 10:05:01.595977 4926 generic.go:334] "Generic (PLEG): container finished" podID="156bc1bd-066f-4071-8d4a-b2124fe381bb" containerID="e9f840bbf118def8a2a798831c802b438ec563ccdefb3f8865aeead33c20991f" exitCode=0 Nov 25 10:05:01 crc kubenswrapper[4926]: I1125 10:05:01.596035 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"156bc1bd-066f-4071-8d4a-b2124fe381bb","Type":"ContainerDied","Data":"e9f840bbf118def8a2a798831c802b438ec563ccdefb3f8865aeead33c20991f"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.608928 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"e7753413-a221-4236-9003-e722280e15bf","Type":"ContainerStarted","Data":"41b941e5171d45e8a6c3087017570e29b07298d53063e1c4b1252f399ce7f96f"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.612287 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkcjx" event={"ID":"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a","Type":"ContainerStarted","Data":"acdc0cd99c078767be67052476fba742ea4cdf406782a172e7f05afff33a6bf4"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.612331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-dkcjx" event={"ID":"dbdb45ad-32a5-4edc-aaa3-d907b5272b8a","Type":"ContainerStarted","Data":"9a63d140b797f9bd22a9d4f54d2da3d1ba4375b4130d52e284b0261710b4fb36"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.612679 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.615401 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2a7b065a-b959-4809-8f8d-1a36b09579c9","Type":"ContainerStarted","Data":"2be40f317412962f996f7988484c59a25b76b07184375e3677e1efc6b8d5cade"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.617752 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"17396f6f-d876-49fe-9294-fe074f9da844","Type":"ContainerStarted","Data":"c3f57dd6bff216a676110cafac172f59fc8ede904bd0fdcf268e8490eaaffc9f"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.619871 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"156bc1bd-066f-4071-8d4a-b2124fe381bb","Type":"ContainerStarted","Data":"a57cc2c8f1bfa5f79bf9b849fee24464f1ccce2a47885dbe25d0f258000ee2c8"} Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.646289 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=7.498636978 podStartE2EDuration="20.646233752s" podCreationTimestamp="2025-11-25 10:04:42 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.255824877 +0000 UTC m=+984.939781304" lastFinishedPulling="2025-11-25 10:05:01.403421651 +0000 UTC m=+998.087378078" observedRunningTime="2025-11-25 10:05:02.642583381 +0000 UTC m=+999.326539818" watchObservedRunningTime="2025-11-25 10:05:02.646233752 +0000 UTC m=+999.330190179" Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.674539 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=19.356183542 podStartE2EDuration="27.674502204s" podCreationTimestamp="2025-11-25 10:04:35 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.175306429 +0000 UTC m=+984.859262856" lastFinishedPulling="2025-11-25 10:04:56.493625091 +0000 UTC m=+993.177581518" observedRunningTime="2025-11-25 10:05:02.667785387 +0000 UTC m=+999.351741824" watchObservedRunningTime="2025-11-25 10:05:02.674502204 +0000 UTC m=+999.358458651" Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.702218 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.558269358 podStartE2EDuration="26.702192961s" podCreationTimestamp="2025-11-25 10:04:36 +0000 UTC" firstStartedPulling="2025-11-25 10:04:47.829976716 +0000 UTC m=+984.513933143" lastFinishedPulling="2025-11-25 10:04:55.973900319 +0000 UTC m=+992.657856746" observedRunningTime="2025-11-25 10:05:02.692959672 +0000 UTC m=+999.376916149" watchObservedRunningTime="2025-11-25 10:05:02.702192961 +0000 UTC m=+999.386149398" Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.726962 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=5.685831308 podStartE2EDuration="17.726924345s" podCreationTimestamp="2025-11-25 10:04:45 +0000 UTC" firstStartedPulling="2025-11-25 10:04:49.352762327 +0000 UTC m=+986.036718754" lastFinishedPulling="2025-11-25 10:05:01.393855364 +0000 UTC m=+998.077811791" observedRunningTime="2025-11-25 10:05:02.725168701 +0000 UTC m=+999.409125158" watchObservedRunningTime="2025-11-25 10:05:02.726924345 +0000 UTC m=+999.410880772" Nov 25 10:05:02 crc kubenswrapper[4926]: I1125 10:05:02.748073 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-dkcjx" podStartSLOduration=12.239041136 podStartE2EDuration="19.748051989s" podCreationTimestamp="2025-11-25 10:04:43 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.343955554 +0000 UTC m=+985.027911981" lastFinishedPulling="2025-11-25 10:04:55.852966397 +0000 UTC m=+992.536922834" observedRunningTime="2025-11-25 10:05:02.745388463 +0000 UTC m=+999.429344940" watchObservedRunningTime="2025-11-25 10:05:02.748051989 +0000 UTC m=+999.432008416" Nov 25 10:05:03 crc kubenswrapper[4926]: I1125 10:05:03.115357 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Nov 25 10:05:03 crc kubenswrapper[4926]: I1125 10:05:03.413114 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:05:03 crc kubenswrapper[4926]: I1125 10:05:03.634278 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:05:03 crc kubenswrapper[4926]: I1125 10:05:03.743873 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:05:03 crc kubenswrapper[4926]: I1125 10:05:03.805406 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-b4qlp"] Nov 25 10:05:03 crc kubenswrapper[4926]: I1125 10:05:03.805870 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerName="dnsmasq-dns" containerID="cri-o://b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7" gracePeriod=10 Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.247587 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.323633 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.324952 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-dns-svc\") pod \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.325285 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-config\") pod \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.325443 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vzdm\" (UniqueName: \"kubernetes.io/projected/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-kube-api-access-6vzdm\") pod \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\" (UID: \"4859cc0d-8bb6-40fc-aa73-f7d2567b314f\") " Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.339062 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-kube-api-access-6vzdm" (OuterVolumeSpecName: "kube-api-access-6vzdm") pod "4859cc0d-8bb6-40fc-aa73-f7d2567b314f" (UID: "4859cc0d-8bb6-40fc-aa73-f7d2567b314f"). InnerVolumeSpecName "kube-api-access-6vzdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.375774 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4859cc0d-8bb6-40fc-aa73-f7d2567b314f" (UID: "4859cc0d-8bb6-40fc-aa73-f7d2567b314f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.381736 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-config" (OuterVolumeSpecName: "config") pod "4859cc0d-8bb6-40fc-aa73-f7d2567b314f" (UID: "4859cc0d-8bb6-40fc-aa73-f7d2567b314f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.419497 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.427733 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.427791 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vzdm\" (UniqueName: \"kubernetes.io/projected/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-kube-api-access-6vzdm\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.427813 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4859cc0d-8bb6-40fc-aa73-f7d2567b314f-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.547415 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.640107 4926 generic.go:334] "Generic (PLEG): container finished" podID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerID="b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7" exitCode=0 Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.640194 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.640253 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" event={"ID":"4859cc0d-8bb6-40fc-aa73-f7d2567b314f","Type":"ContainerDied","Data":"b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7"} Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.640286 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-666b6646f7-b4qlp" event={"ID":"4859cc0d-8bb6-40fc-aa73-f7d2567b314f","Type":"ContainerDied","Data":"44b68d190c018715f762faea0ed9bb2b493eec326cb73d70a123a114f46ff512"} Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.640306 4926 scope.go:117] "RemoveContainer" containerID="b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.641159 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.665629 4926 scope.go:117] "RemoveContainer" containerID="4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.671862 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-b4qlp"] Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.678003 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-666b6646f7-b4qlp"] Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.681370 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.702357 4926 scope.go:117] "RemoveContainer" containerID="b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7" Nov 25 10:05:04 crc kubenswrapper[4926]: E1125 10:05:04.703026 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7\": container with ID starting with b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7 not found: ID does not exist" containerID="b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.703083 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7"} err="failed to get container status \"b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7\": rpc error: code = NotFound desc = could not find container \"b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7\": container with ID starting with b14308c0c4fbc77e45812c9356c7825e37b20a56c75b1a1c12422a6afaded0d7 not found: ID does not exist" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.703119 4926 scope.go:117] "RemoveContainer" containerID="4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924" Nov 25 10:05:04 crc kubenswrapper[4926]: E1125 10:05:04.703809 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924\": container with ID starting with 4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924 not found: ID does not exist" containerID="4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.703866 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924"} err="failed to get container status \"4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924\": rpc error: code = NotFound desc = could not find container \"4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924\": container with ID starting with 4bcb71d825eaac57c4d991b8c0324fe4db98d796a8fedd0f49bdcb72e8479924 not found: ID does not exist" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.972378 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9kfqp"] Nov 25 10:05:04 crc kubenswrapper[4926]: E1125 10:05:04.972926 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerName="init" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.972986 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerName="init" Nov 25 10:05:04 crc kubenswrapper[4926]: E1125 10:05:04.973023 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerName="dnsmasq-dns" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.973036 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerName="dnsmasq-dns" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.973273 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" containerName="dnsmasq-dns" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.975423 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.983727 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9kfqp"] Nov 25 10:05:04 crc kubenswrapper[4926]: I1125 10:05:04.985955 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.037200 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-config\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.037548 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.037660 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mspwd\" (UniqueName: \"kubernetes.io/projected/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-kube-api-access-mspwd\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.037714 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.069961 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-5xppx"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.071085 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.074412 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.085675 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5xppx"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139416 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1386c3-686d-4f5f-85b1-2f721ed3be80-combined-ca-bundle\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139469 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5mdm\" (UniqueName: \"kubernetes.io/projected/1a1386c3-686d-4f5f-85b1-2f721ed3be80-kube-api-access-h5mdm\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139568 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-config\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139597 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139622 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1386c3-686d-4f5f-85b1-2f721ed3be80-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139668 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a1386c3-686d-4f5f-85b1-2f721ed3be80-config\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1a1386c3-686d-4f5f-85b1-2f721ed3be80-ovs-rundir\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139736 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1a1386c3-686d-4f5f-85b1-2f721ed3be80-ovn-rundir\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139778 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mspwd\" (UniqueName: \"kubernetes.io/projected/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-kube-api-access-mspwd\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.139799 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.141553 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-config\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.141571 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-dns-svc\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.142330 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-ovsdbserver-nb\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.160913 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mspwd\" (UniqueName: \"kubernetes.io/projected/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-kube-api-access-mspwd\") pod \"dnsmasq-dns-5bf47b49b7-9kfqp\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.241283 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1a1386c3-686d-4f5f-85b1-2f721ed3be80-ovs-rundir\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.241331 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1a1386c3-686d-4f5f-85b1-2f721ed3be80-ovn-rundir\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.241378 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5mdm\" (UniqueName: \"kubernetes.io/projected/1a1386c3-686d-4f5f-85b1-2f721ed3be80-kube-api-access-h5mdm\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.241398 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1386c3-686d-4f5f-85b1-2f721ed3be80-combined-ca-bundle\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.241459 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1386c3-686d-4f5f-85b1-2f721ed3be80-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.241491 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a1386c3-686d-4f5f-85b1-2f721ed3be80-config\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.242265 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a1386c3-686d-4f5f-85b1-2f721ed3be80-config\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.242512 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/1a1386c3-686d-4f5f-85b1-2f721ed3be80-ovs-rundir\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.242595 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/1a1386c3-686d-4f5f-85b1-2f721ed3be80-ovn-rundir\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.247704 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/1a1386c3-686d-4f5f-85b1-2f721ed3be80-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.249983 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a1386c3-686d-4f5f-85b1-2f721ed3be80-combined-ca-bundle\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.275434 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5mdm\" (UniqueName: \"kubernetes.io/projected/1a1386c3-686d-4f5f-85b1-2f721ed3be80-kube-api-access-h5mdm\") pod \"ovn-controller-metrics-5xppx\" (UID: \"1a1386c3-686d-4f5f-85b1-2f721ed3be80\") " pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.294772 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.323304 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.381550 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9kfqp"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.392093 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-5xppx" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.404184 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-lfmd2"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.405902 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.408117 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.415734 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.428573 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lfmd2"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.566256 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-config\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.566746 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.566793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.566850 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-dns-svc\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.566899 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqk68\" (UniqueName: \"kubernetes.io/projected/9b82a532-0d0e-4cf3-892c-540122644a0e-kube-api-access-hqk68\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.670145 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-dns-svc\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.670226 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqk68\" (UniqueName: \"kubernetes.io/projected/9b82a532-0d0e-4cf3-892c-540122644a0e-kube-api-access-hqk68\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.670299 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-config\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.670340 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.670374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.671560 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-dns-svc\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.671614 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-config\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.671728 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.671785 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.685661 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9kfqp"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.692772 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqk68\" (UniqueName: \"kubernetes.io/projected/9b82a532-0d0e-4cf3-892c-540122644a0e-kube-api-access-hqk68\") pod \"dnsmasq-dns-8554648995-lfmd2\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.717315 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.739533 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.973909 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.976019 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.980423 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.983401 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.983596 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.987796 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-p77xz" Nov 25 10:05:05 crc kubenswrapper[4926]: I1125 10:05:05.993009 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.033511 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4859cc0d-8bb6-40fc-aa73-f7d2567b314f" path="/var/lib/kubelet/pods/4859cc0d-8bb6-40fc-aa73-f7d2567b314f/volumes" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.034838 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-5xppx"] Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077115 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077176 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077288 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/513417d4-ae93-44eb-a5e9-f16094bd3330-scripts\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077305 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/513417d4-ae93-44eb-a5e9-f16094bd3330-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077342 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lbv4\" (UniqueName: \"kubernetes.io/projected/513417d4-ae93-44eb-a5e9-f16094bd3330-kube-api-access-9lbv4\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077374 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.077426 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/513417d4-ae93-44eb-a5e9-f16094bd3330-config\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179247 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/513417d4-ae93-44eb-a5e9-f16094bd3330-scripts\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179500 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/513417d4-ae93-44eb-a5e9-f16094bd3330-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179529 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lbv4\" (UniqueName: \"kubernetes.io/projected/513417d4-ae93-44eb-a5e9-f16094bd3330-kube-api-access-9lbv4\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179569 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179592 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/513417d4-ae93-44eb-a5e9-f16094bd3330-config\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.179645 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.181280 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/513417d4-ae93-44eb-a5e9-f16094bd3330-scripts\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.181446 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/513417d4-ae93-44eb-a5e9-f16094bd3330-config\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.181666 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/513417d4-ae93-44eb-a5e9-f16094bd3330-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.186119 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.186971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.187133 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/513417d4-ae93-44eb-a5e9-f16094bd3330-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.199442 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lbv4\" (UniqueName: \"kubernetes.io/projected/513417d4-ae93-44eb-a5e9-f16094bd3330-kube-api-access-9lbv4\") pod \"ovn-northd-0\" (UID: \"513417d4-ae93-44eb-a5e9-f16094bd3330\") " pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.317244 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.349015 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lfmd2"] Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.509494 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.509872 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.685925 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5xppx" event={"ID":"1a1386c3-686d-4f5f-85b1-2f721ed3be80","Type":"ContainerStarted","Data":"89f173cfdbab89ffe84bff702b701747dd8f7281f149d7729884c6ee0ccca236"} Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.698984 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" event={"ID":"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c","Type":"ContainerStarted","Data":"49d1e523098948952aa88a26e5614b4d94e5860d5e4d83435709b8b12eb57fc8"} Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.710592 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lfmd2" event={"ID":"9b82a532-0d0e-4cf3-892c-540122644a0e","Type":"ContainerStarted","Data":"cc9df3c5be74e3566c485f35398453098bb02a196f2a9ecbe343ca47cbc9b9ce"} Nov 25 10:05:06 crc kubenswrapper[4926]: I1125 10:05:06.843965 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Nov 25 10:05:07 crc kubenswrapper[4926]: I1125 10:05:07.721154 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"513417d4-ae93-44eb-a5e9-f16094bd3330","Type":"ContainerStarted","Data":"abc50cdcd347de6bc80a8d022b1d0bc2d9cb969f144b1114e564ed1ba1ecde31"} Nov 25 10:05:07 crc kubenswrapper[4926]: I1125 10:05:07.803562 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Nov 25 10:05:07 crc kubenswrapper[4926]: I1125 10:05:07.803649 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Nov 25 10:05:09 crc kubenswrapper[4926]: I1125 10:05:09.740278 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:05:10 crc kubenswrapper[4926]: I1125 10:05:10.745319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" event={"ID":"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c","Type":"ContainerStarted","Data":"349b7f5e4dbf4b39a8c521bd7a7c706e4b1a4394fafaef3d030cc399d1d0071d"} Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.655557 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.733488 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.753319 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-5xppx" event={"ID":"1a1386c3-686d-4f5f-85b1-2f721ed3be80","Type":"ContainerStarted","Data":"dc782a39292df64eb5a20ff51af45112df12b39d1b62cfe4957393ce5da317ff"} Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.755981 4926 generic.go:334] "Generic (PLEG): container finished" podID="61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" containerID="349b7f5e4dbf4b39a8c521bd7a7c706e4b1a4394fafaef3d030cc399d1d0071d" exitCode=0 Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.756028 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" event={"ID":"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c","Type":"ContainerDied","Data":"349b7f5e4dbf4b39a8c521bd7a7c706e4b1a4394fafaef3d030cc399d1d0071d"} Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.758061 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerID="3a896daa993688f15b9f07a8138bf33cdd5f905689a38a5d83bf483079a3ce27" exitCode=0 Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.759294 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lfmd2" event={"ID":"9b82a532-0d0e-4cf3-892c-540122644a0e","Type":"ContainerDied","Data":"3a896daa993688f15b9f07a8138bf33cdd5f905689a38a5d83bf483079a3ce27"} Nov 25 10:05:11 crc kubenswrapper[4926]: I1125 10:05:11.790487 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-5xppx" podStartSLOduration=6.790464169 podStartE2EDuration="6.790464169s" podCreationTimestamp="2025-11-25 10:05:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:05:11.786857189 +0000 UTC m=+1008.470813626" watchObservedRunningTime="2025-11-25 10:05:11.790464169 +0000 UTC m=+1008.474420616" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.165670 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.296514 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mspwd\" (UniqueName: \"kubernetes.io/projected/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-kube-api-access-mspwd\") pod \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.296666 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-ovsdbserver-nb\") pod \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.296720 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-dns-svc\") pod \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.296853 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-config\") pod \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\" (UID: \"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c\") " Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.302026 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-kube-api-access-mspwd" (OuterVolumeSpecName: "kube-api-access-mspwd") pod "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" (UID: "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c"). InnerVolumeSpecName "kube-api-access-mspwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.320387 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" (UID: "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.321959 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-config" (OuterVolumeSpecName: "config") pod "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" (UID: "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.324158 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" (UID: "61fa940f-dc50-4ee6-abc7-aaa0b0ca083c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.398526 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.398572 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mspwd\" (UniqueName: \"kubernetes.io/projected/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-kube-api-access-mspwd\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.398586 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.398596 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.771333 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lfmd2" event={"ID":"9b82a532-0d0e-4cf3-892c-540122644a0e","Type":"ContainerStarted","Data":"50538bb73cfc013a33e6a8e278d0a37af22444c3674886483ef937af8d9b7bb4"} Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.771842 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.774495 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" event={"ID":"61fa940f-dc50-4ee6-abc7-aaa0b0ca083c","Type":"ContainerDied","Data":"49d1e523098948952aa88a26e5614b4d94e5860d5e4d83435709b8b12eb57fc8"} Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.774541 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bf47b49b7-9kfqp" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.774568 4926 scope.go:117] "RemoveContainer" containerID="349b7f5e4dbf4b39a8c521bd7a7c706e4b1a4394fafaef3d030cc399d1d0071d" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.786152 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"513417d4-ae93-44eb-a5e9-f16094bd3330","Type":"ContainerStarted","Data":"d7db8b7256ab42d0866afbc2a0d451342ce2e9ad18b87a1d090dbc0fed86d1fb"} Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.786221 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"513417d4-ae93-44eb-a5e9-f16094bd3330","Type":"ContainerStarted","Data":"d3a1b8177ec9e55c7ec5b816973589f87a64c18a937628e4c5b6a7f4e41de41b"} Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.786259 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.801303 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-lfmd2" podStartSLOduration=7.801285302 podStartE2EDuration="7.801285302s" podCreationTimestamp="2025-11-25 10:05:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:05:12.795073567 +0000 UTC m=+1009.479029994" watchObservedRunningTime="2025-11-25 10:05:12.801285302 +0000 UTC m=+1009.485241729" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.824712 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=3.08933195 podStartE2EDuration="7.824683602s" podCreationTimestamp="2025-11-25 10:05:05 +0000 UTC" firstStartedPulling="2025-11-25 10:05:06.846804036 +0000 UTC m=+1003.530760463" lastFinishedPulling="2025-11-25 10:05:11.582155668 +0000 UTC m=+1008.266112115" observedRunningTime="2025-11-25 10:05:12.818314814 +0000 UTC m=+1009.502271241" watchObservedRunningTime="2025-11-25 10:05:12.824683602 +0000 UTC m=+1009.508640029" Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.864684 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9kfqp"] Nov 25 10:05:12 crc kubenswrapper[4926]: I1125 10:05:12.873271 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bf47b49b7-9kfqp"] Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.387341 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-pmc8k"] Nov 25 10:05:13 crc kubenswrapper[4926]: E1125 10:05:13.387787 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" containerName="init" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.387810 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" containerName="init" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.388042 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" containerName="init" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.388743 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.393520 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-298a-account-create-fpwbz"] Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.394594 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.396021 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.406316 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pmc8k"] Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.432635 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-298a-account-create-fpwbz"] Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.518335 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/479a95d6-7f72-4322-929a-d124f7722c40-operator-scripts\") pod \"glance-db-create-pmc8k\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.518409 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgdqz\" (UniqueName: \"kubernetes.io/projected/230dd925-418b-4bc1-b62f-b70befdd2a59-kube-api-access-qgdqz\") pod \"glance-298a-account-create-fpwbz\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.518456 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm84v\" (UniqueName: \"kubernetes.io/projected/479a95d6-7f72-4322-929a-d124f7722c40-kube-api-access-xm84v\") pod \"glance-db-create-pmc8k\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.518483 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/230dd925-418b-4bc1-b62f-b70befdd2a59-operator-scripts\") pod \"glance-298a-account-create-fpwbz\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.620033 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/479a95d6-7f72-4322-929a-d124f7722c40-operator-scripts\") pod \"glance-db-create-pmc8k\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.620098 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgdqz\" (UniqueName: \"kubernetes.io/projected/230dd925-418b-4bc1-b62f-b70befdd2a59-kube-api-access-qgdqz\") pod \"glance-298a-account-create-fpwbz\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.620143 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm84v\" (UniqueName: \"kubernetes.io/projected/479a95d6-7f72-4322-929a-d124f7722c40-kube-api-access-xm84v\") pod \"glance-db-create-pmc8k\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.620180 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/230dd925-418b-4bc1-b62f-b70befdd2a59-operator-scripts\") pod \"glance-298a-account-create-fpwbz\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.621057 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/479a95d6-7f72-4322-929a-d124f7722c40-operator-scripts\") pod \"glance-db-create-pmc8k\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.621220 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/230dd925-418b-4bc1-b62f-b70befdd2a59-operator-scripts\") pod \"glance-298a-account-create-fpwbz\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.636856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgdqz\" (UniqueName: \"kubernetes.io/projected/230dd925-418b-4bc1-b62f-b70befdd2a59-kube-api-access-qgdqz\") pod \"glance-298a-account-create-fpwbz\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.641528 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm84v\" (UniqueName: \"kubernetes.io/projected/479a95d6-7f72-4322-929a-d124f7722c40-kube-api-access-xm84v\") pod \"glance-db-create-pmc8k\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.711400 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.733340 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:13 crc kubenswrapper[4926]: I1125 10:05:13.903518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.043911 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="61fa940f-dc50-4ee6-abc7-aaa0b0ca083c" path="/var/lib/kubelet/pods/61fa940f-dc50-4ee6-abc7-aaa0b0ca083c/volumes" Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.148353 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.193630 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-pmc8k"] Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.198090 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-298a-account-create-fpwbz"] Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.802611 4926 generic.go:334] "Generic (PLEG): container finished" podID="230dd925-418b-4bc1-b62f-b70befdd2a59" containerID="72f9706efbd33dcb0f06a30f151666de3d9ea9209b67403a4ea7d476080f813c" exitCode=0 Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.802778 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-298a-account-create-fpwbz" event={"ID":"230dd925-418b-4bc1-b62f-b70befdd2a59","Type":"ContainerDied","Data":"72f9706efbd33dcb0f06a30f151666de3d9ea9209b67403a4ea7d476080f813c"} Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.803009 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-298a-account-create-fpwbz" event={"ID":"230dd925-418b-4bc1-b62f-b70befdd2a59","Type":"ContainerStarted","Data":"c084eb2ab8260eea33ee090a9baf95544b9a1815588988280f48b46b0f13bac3"} Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.804966 4926 generic.go:334] "Generic (PLEG): container finished" podID="479a95d6-7f72-4322-929a-d124f7722c40" containerID="df8e3130e3f3beaa8b38fdb49b0e47a61cb1d15fc72b2c5aacc1df874dfbe2b4" exitCode=0 Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.805102 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmc8k" event={"ID":"479a95d6-7f72-4322-929a-d124f7722c40","Type":"ContainerDied","Data":"df8e3130e3f3beaa8b38fdb49b0e47a61cb1d15fc72b2c5aacc1df874dfbe2b4"} Nov 25 10:05:14 crc kubenswrapper[4926]: I1125 10:05:14.805158 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmc8k" event={"ID":"479a95d6-7f72-4322-929a-d124f7722c40","Type":"ContainerStarted","Data":"47b33650562dbf559eb55cd24c0ebdcbadcf128b2f03aa44d146ad3b32ab87b8"} Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.257646 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.265530 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.392386 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm84v\" (UniqueName: \"kubernetes.io/projected/479a95d6-7f72-4322-929a-d124f7722c40-kube-api-access-xm84v\") pod \"479a95d6-7f72-4322-929a-d124f7722c40\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.392443 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/479a95d6-7f72-4322-929a-d124f7722c40-operator-scripts\") pod \"479a95d6-7f72-4322-929a-d124f7722c40\" (UID: \"479a95d6-7f72-4322-929a-d124f7722c40\") " Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.392503 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/230dd925-418b-4bc1-b62f-b70befdd2a59-operator-scripts\") pod \"230dd925-418b-4bc1-b62f-b70befdd2a59\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.392518 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgdqz\" (UniqueName: \"kubernetes.io/projected/230dd925-418b-4bc1-b62f-b70befdd2a59-kube-api-access-qgdqz\") pod \"230dd925-418b-4bc1-b62f-b70befdd2a59\" (UID: \"230dd925-418b-4bc1-b62f-b70befdd2a59\") " Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.393627 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/230dd925-418b-4bc1-b62f-b70befdd2a59-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "230dd925-418b-4bc1-b62f-b70befdd2a59" (UID: "230dd925-418b-4bc1-b62f-b70befdd2a59"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.393630 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/479a95d6-7f72-4322-929a-d124f7722c40-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "479a95d6-7f72-4322-929a-d124f7722c40" (UID: "479a95d6-7f72-4322-929a-d124f7722c40"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.398222 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/230dd925-418b-4bc1-b62f-b70befdd2a59-kube-api-access-qgdqz" (OuterVolumeSpecName: "kube-api-access-qgdqz") pod "230dd925-418b-4bc1-b62f-b70befdd2a59" (UID: "230dd925-418b-4bc1-b62f-b70befdd2a59"). InnerVolumeSpecName "kube-api-access-qgdqz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.399045 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/479a95d6-7f72-4322-929a-d124f7722c40-kube-api-access-xm84v" (OuterVolumeSpecName: "kube-api-access-xm84v") pod "479a95d6-7f72-4322-929a-d124f7722c40" (UID: "479a95d6-7f72-4322-929a-d124f7722c40"). InnerVolumeSpecName "kube-api-access-xm84v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.494769 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/230dd925-418b-4bc1-b62f-b70befdd2a59-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.494802 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qgdqz\" (UniqueName: \"kubernetes.io/projected/230dd925-418b-4bc1-b62f-b70befdd2a59-kube-api-access-qgdqz\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.494814 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm84v\" (UniqueName: \"kubernetes.io/projected/479a95d6-7f72-4322-929a-d124f7722c40-kube-api-access-xm84v\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.494823 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/479a95d6-7f72-4322-929a-d124f7722c40-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.828678 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-298a-account-create-fpwbz" event={"ID":"230dd925-418b-4bc1-b62f-b70befdd2a59","Type":"ContainerDied","Data":"c084eb2ab8260eea33ee090a9baf95544b9a1815588988280f48b46b0f13bac3"} Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.828741 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c084eb2ab8260eea33ee090a9baf95544b9a1815588988280f48b46b0f13bac3" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.828704 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-298a-account-create-fpwbz" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.830707 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-pmc8k" event={"ID":"479a95d6-7f72-4322-929a-d124f7722c40","Type":"ContainerDied","Data":"47b33650562dbf559eb55cd24c0ebdcbadcf128b2f03aa44d146ad3b32ab87b8"} Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.830734 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47b33650562dbf559eb55cd24c0ebdcbadcf128b2f03aa44d146ad3b32ab87b8" Nov 25 10:05:16 crc kubenswrapper[4926]: I1125 10:05:16.830767 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-pmc8k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.680726 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9vf4k"] Nov 25 10:05:17 crc kubenswrapper[4926]: E1125 10:05:17.681496 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="230dd925-418b-4bc1-b62f-b70befdd2a59" containerName="mariadb-account-create" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.681517 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="230dd925-418b-4bc1-b62f-b70befdd2a59" containerName="mariadb-account-create" Nov 25 10:05:17 crc kubenswrapper[4926]: E1125 10:05:17.681543 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="479a95d6-7f72-4322-929a-d124f7722c40" containerName="mariadb-database-create" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.681550 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="479a95d6-7f72-4322-929a-d124f7722c40" containerName="mariadb-database-create" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.681748 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="479a95d6-7f72-4322-929a-d124f7722c40" containerName="mariadb-database-create" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.681772 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="230dd925-418b-4bc1-b62f-b70befdd2a59" containerName="mariadb-account-create" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.682607 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.686360 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9vf4k"] Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.793961 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-b21d-account-create-cbzxv"] Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.795377 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.799245 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.814100 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-operator-scripts\") pod \"keystone-db-create-9vf4k\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.814233 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlj8\" (UniqueName: \"kubernetes.io/projected/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-kube-api-access-thlj8\") pod \"keystone-db-create-9vf4k\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.824284 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b21d-account-create-cbzxv"] Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.917029 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bac7de-ff6d-4b5f-85c0-c3653855f844-operator-scripts\") pod \"keystone-b21d-account-create-cbzxv\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.917124 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-operator-scripts\") pod \"keystone-db-create-9vf4k\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.917180 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n65w6\" (UniqueName: \"kubernetes.io/projected/f2bac7de-ff6d-4b5f-85c0-c3653855f844-kube-api-access-n65w6\") pod \"keystone-b21d-account-create-cbzxv\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.917269 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlj8\" (UniqueName: \"kubernetes.io/projected/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-kube-api-access-thlj8\") pod \"keystone-db-create-9vf4k\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.918974 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-operator-scripts\") pod \"keystone-db-create-9vf4k\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.943990 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlj8\" (UniqueName: \"kubernetes.io/projected/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-kube-api-access-thlj8\") pod \"keystone-db-create-9vf4k\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.995520 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-b544s"] Nov 25 10:05:17 crc kubenswrapper[4926]: I1125 10:05:17.996820 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.018767 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.019205 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n65w6\" (UniqueName: \"kubernetes.io/projected/f2bac7de-ff6d-4b5f-85c0-c3653855f844-kube-api-access-n65w6\") pod \"keystone-b21d-account-create-cbzxv\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.019303 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bac7de-ff6d-4b5f-85c0-c3653855f844-operator-scripts\") pod \"keystone-b21d-account-create-cbzxv\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.020003 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bac7de-ff6d-4b5f-85c0-c3653855f844-operator-scripts\") pod \"keystone-b21d-account-create-cbzxv\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.032859 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b544s"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.045795 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n65w6\" (UniqueName: \"kubernetes.io/projected/f2bac7de-ff6d-4b5f-85c0-c3653855f844-kube-api-access-n65w6\") pod \"keystone-b21d-account-create-cbzxv\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.111551 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-fa79-account-create-c9vvp"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.112927 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.115203 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.117270 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.122833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6n6r\" (UniqueName: \"kubernetes.io/projected/b24c1a1f-5324-4699-90a5-b539c855acb6-kube-api-access-g6n6r\") pod \"placement-db-create-b544s\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.123016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c1a1f-5324-4699-90a5-b539c855acb6-operator-scripts\") pod \"placement-db-create-b544s\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.163230 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fa79-account-create-c9vvp"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.224362 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-operator-scripts\") pod \"placement-fa79-account-create-c9vvp\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.224435 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c1a1f-5324-4699-90a5-b539c855acb6-operator-scripts\") pod \"placement-db-create-b544s\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.224523 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqxxq\" (UniqueName: \"kubernetes.io/projected/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-kube-api-access-cqxxq\") pod \"placement-fa79-account-create-c9vvp\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.224580 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6n6r\" (UniqueName: \"kubernetes.io/projected/b24c1a1f-5324-4699-90a5-b539c855acb6-kube-api-access-g6n6r\") pod \"placement-db-create-b544s\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.225354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c1a1f-5324-4699-90a5-b539c855acb6-operator-scripts\") pod \"placement-db-create-b544s\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.250372 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6n6r\" (UniqueName: \"kubernetes.io/projected/b24c1a1f-5324-4699-90a5-b539c855acb6-kube-api-access-g6n6r\") pod \"placement-db-create-b544s\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.312676 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b544s" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.325732 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqxxq\" (UniqueName: \"kubernetes.io/projected/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-kube-api-access-cqxxq\") pod \"placement-fa79-account-create-c9vvp\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.325849 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-operator-scripts\") pod \"placement-fa79-account-create-c9vvp\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.326705 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-operator-scripts\") pod \"placement-fa79-account-create-c9vvp\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.347381 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqxxq\" (UniqueName: \"kubernetes.io/projected/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-kube-api-access-cqxxq\") pod \"placement-fa79-account-create-c9vvp\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.518987 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.531797 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9vf4k"] Nov 25 10:05:18 crc kubenswrapper[4926]: W1125 10:05:18.548670 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc48a5eda_1ab4_41d5_9a82_0b6e22a2bef0.slice/crio-5d6f4602259a73d3250a9650db27d8b68b124fdfcf17abba5719cb066118fd7b WatchSource:0}: Error finding container 5d6f4602259a73d3250a9650db27d8b68b124fdfcf17abba5719cb066118fd7b: Status 404 returned error can't find the container with id 5d6f4602259a73d3250a9650db27d8b68b124fdfcf17abba5719cb066118fd7b Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.674904 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-b21d-account-create-cbzxv"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.721386 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-grrv7"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.726105 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.732440 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-grrv7"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.735117 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.735164 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-9pvwn" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.772775 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-b544s"] Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.835309 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-db-sync-config-data\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.835388 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-config-data\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.835474 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpnh7\" (UniqueName: \"kubernetes.io/projected/ceab1d9b-89a4-412c-b720-b46193249d81-kube-api-access-fpnh7\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.835616 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-combined-ca-bundle\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.864672 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b544s" event={"ID":"b24c1a1f-5324-4699-90a5-b539c855acb6","Type":"ContainerStarted","Data":"234fdbdd6b1b2c1aaa1b61b22ff83450b84c90021fb7afc5d7b2fa0f33b09895"} Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.870151 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b21d-account-create-cbzxv" event={"ID":"f2bac7de-ff6d-4b5f-85c0-c3653855f844","Type":"ContainerStarted","Data":"85def9b7fd3bed2459efd73524d017099a937e7a39bd4ace2b1b5b1538e46d60"} Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.872306 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9vf4k" event={"ID":"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0","Type":"ContainerStarted","Data":"6a6c361b7ee2665024d77674ec3b601fa009b3f406f54f5843f44f46f20a1711"} Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.872332 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9vf4k" event={"ID":"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0","Type":"ContainerStarted","Data":"5d6f4602259a73d3250a9650db27d8b68b124fdfcf17abba5719cb066118fd7b"} Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.892480 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-9vf4k" podStartSLOduration=1.892456107 podStartE2EDuration="1.892456107s" podCreationTimestamp="2025-11-25 10:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:05:18.890390406 +0000 UTC m=+1015.574346833" watchObservedRunningTime="2025-11-25 10:05:18.892456107 +0000 UTC m=+1015.576412534" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.937614 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-db-sync-config-data\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.937670 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-config-data\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.937769 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpnh7\" (UniqueName: \"kubernetes.io/projected/ceab1d9b-89a4-412c-b720-b46193249d81-kube-api-access-fpnh7\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.938283 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-combined-ca-bundle\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.943485 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-config-data\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.945196 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-db-sync-config-data\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.946645 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-combined-ca-bundle\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:18 crc kubenswrapper[4926]: I1125 10:05:18.958990 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpnh7\" (UniqueName: \"kubernetes.io/projected/ceab1d9b-89a4-412c-b720-b46193249d81-kube-api-access-fpnh7\") pod \"glance-db-sync-grrv7\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.014482 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-fa79-account-create-c9vvp"] Nov 25 10:05:19 crc kubenswrapper[4926]: W1125 10:05:19.020326 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod109f1cf1_f016_438b_b2b9_2eaabf3ba1c9.slice/crio-de5a42967d9a61cc7cc33e99685e1fdfea264c67957bb998e3e5a09a7c94aea4 WatchSource:0}: Error finding container de5a42967d9a61cc7cc33e99685e1fdfea264c67957bb998e3e5a09a7c94aea4: Status 404 returned error can't find the container with id de5a42967d9a61cc7cc33e99685e1fdfea264c67957bb998e3e5a09a7c94aea4 Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.054857 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:19 crc kubenswrapper[4926]: W1125 10:05:19.558014 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podceab1d9b_89a4_412c_b720_b46193249d81.slice/crio-2553640bff97827ea34a7e7913aa97034385ea5658dc747b75b56e6fad61ee36 WatchSource:0}: Error finding container 2553640bff97827ea34a7e7913aa97034385ea5658dc747b75b56e6fad61ee36: Status 404 returned error can't find the container with id 2553640bff97827ea34a7e7913aa97034385ea5658dc747b75b56e6fad61ee36 Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.562477 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-grrv7"] Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.884069 4926 generic.go:334] "Generic (PLEG): container finished" podID="c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" containerID="6a6c361b7ee2665024d77674ec3b601fa009b3f406f54f5843f44f46f20a1711" exitCode=0 Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.884776 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9vf4k" event={"ID":"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0","Type":"ContainerDied","Data":"6a6c361b7ee2665024d77674ec3b601fa009b3f406f54f5843f44f46f20a1711"} Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.886060 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-grrv7" event={"ID":"ceab1d9b-89a4-412c-b720-b46193249d81","Type":"ContainerStarted","Data":"2553640bff97827ea34a7e7913aa97034385ea5658dc747b75b56e6fad61ee36"} Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.888015 4926 generic.go:334] "Generic (PLEG): container finished" podID="109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" containerID="e8e58cf43c7ecf87b491b558284ddfb60380a8ba9ca0087dac53819159571b5b" exitCode=0 Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.888088 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fa79-account-create-c9vvp" event={"ID":"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9","Type":"ContainerDied","Data":"e8e58cf43c7ecf87b491b558284ddfb60380a8ba9ca0087dac53819159571b5b"} Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.888113 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fa79-account-create-c9vvp" event={"ID":"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9","Type":"ContainerStarted","Data":"de5a42967d9a61cc7cc33e99685e1fdfea264c67957bb998e3e5a09a7c94aea4"} Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.890202 4926 generic.go:334] "Generic (PLEG): container finished" podID="b24c1a1f-5324-4699-90a5-b539c855acb6" containerID="6b1ced5690cceae03aea9163438b50de336a67bdbc21f305a8a4946b9a26fcc0" exitCode=0 Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.890296 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b544s" event={"ID":"b24c1a1f-5324-4699-90a5-b539c855acb6","Type":"ContainerDied","Data":"6b1ced5690cceae03aea9163438b50de336a67bdbc21f305a8a4946b9a26fcc0"} Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.894337 4926 generic.go:334] "Generic (PLEG): container finished" podID="f2bac7de-ff6d-4b5f-85c0-c3653855f844" containerID="e08ffb87f1516f45eb4fb713ffb56a32d79a6a754fb21099998e7f5b59637f67" exitCode=0 Nov 25 10:05:19 crc kubenswrapper[4926]: I1125 10:05:19.894386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b21d-account-create-cbzxv" event={"ID":"f2bac7de-ff6d-4b5f-85c0-c3653855f844","Type":"ContainerDied","Data":"e08ffb87f1516f45eb4fb713ffb56a32d79a6a754fb21099998e7f5b59637f67"} Nov 25 10:05:20 crc kubenswrapper[4926]: I1125 10:05:20.741113 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:20 crc kubenswrapper[4926]: I1125 10:05:20.813629 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhgqq"] Nov 25 10:05:20 crc kubenswrapper[4926]: I1125 10:05:20.814021 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" containerName="dnsmasq-dns" containerID="cri-o://7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0" gracePeriod=10 Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.413772 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.492081 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.614407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqxxq\" (UniqueName: \"kubernetes.io/projected/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-kube-api-access-cqxxq\") pod \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.614480 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-operator-scripts\") pod \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\" (UID: \"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.615282 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" (UID: "109f1cf1-f016-438b-b2b9-2eaabf3ba1c9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.646353 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-kube-api-access-cqxxq" (OuterVolumeSpecName: "kube-api-access-cqxxq") pod "109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" (UID: "109f1cf1-f016-438b-b2b9-2eaabf3ba1c9"). InnerVolumeSpecName "kube-api-access-cqxxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.715860 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqxxq\" (UniqueName: \"kubernetes.io/projected/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-kube-api-access-cqxxq\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.715905 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.740228 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b544s" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.750772 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.767277 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.794997 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.819016 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c1a1f-5324-4699-90a5-b539c855acb6-operator-scripts\") pod \"b24c1a1f-5324-4699-90a5-b539c855acb6\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.819331 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-operator-scripts\") pod \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.819460 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n65w6\" (UniqueName: \"kubernetes.io/projected/f2bac7de-ff6d-4b5f-85c0-c3653855f844-kube-api-access-n65w6\") pod \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.819591 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thlj8\" (UniqueName: \"kubernetes.io/projected/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-kube-api-access-thlj8\") pod \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\" (UID: \"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.819689 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6n6r\" (UniqueName: \"kubernetes.io/projected/b24c1a1f-5324-4699-90a5-b539c855acb6-kube-api-access-g6n6r\") pod \"b24c1a1f-5324-4699-90a5-b539c855acb6\" (UID: \"b24c1a1f-5324-4699-90a5-b539c855acb6\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.819998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bac7de-ff6d-4b5f-85c0-c3653855f844-operator-scripts\") pod \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\" (UID: \"f2bac7de-ff6d-4b5f-85c0-c3653855f844\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.822514 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b24c1a1f-5324-4699-90a5-b539c855acb6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b24c1a1f-5324-4699-90a5-b539c855acb6" (UID: "b24c1a1f-5324-4699-90a5-b539c855acb6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.823129 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b24c1a1f-5324-4699-90a5-b539c855acb6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.823718 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2bac7de-ff6d-4b5f-85c0-c3653855f844-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f2bac7de-ff6d-4b5f-85c0-c3653855f844" (UID: "f2bac7de-ff6d-4b5f-85c0-c3653855f844"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.823846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" (UID: "c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.826485 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2bac7de-ff6d-4b5f-85c0-c3653855f844-kube-api-access-n65w6" (OuterVolumeSpecName: "kube-api-access-n65w6") pod "f2bac7de-ff6d-4b5f-85c0-c3653855f844" (UID: "f2bac7de-ff6d-4b5f-85c0-c3653855f844"). InnerVolumeSpecName "kube-api-access-n65w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.827758 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-kube-api-access-thlj8" (OuterVolumeSpecName: "kube-api-access-thlj8") pod "c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" (UID: "c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0"). InnerVolumeSpecName "kube-api-access-thlj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.837905 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b24c1a1f-5324-4699-90a5-b539c855acb6-kube-api-access-g6n6r" (OuterVolumeSpecName: "kube-api-access-g6n6r") pod "b24c1a1f-5324-4699-90a5-b539c855acb6" (UID: "b24c1a1f-5324-4699-90a5-b539c855acb6"). InnerVolumeSpecName "kube-api-access-g6n6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.921897 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-fa79-account-create-c9vvp" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.922077 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-fa79-account-create-c9vvp" event={"ID":"109f1cf1-f016-438b-b2b9-2eaabf3ba1c9","Type":"ContainerDied","Data":"de5a42967d9a61cc7cc33e99685e1fdfea264c67957bb998e3e5a09a7c94aea4"} Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.922144 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de5a42967d9a61cc7cc33e99685e1fdfea264c67957bb998e3e5a09a7c94aea4" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.923880 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvc84\" (UniqueName: \"kubernetes.io/projected/43ea3e47-5f76-4668-84d7-dae46d461369-kube-api-access-bvc84\") pod \"43ea3e47-5f76-4668-84d7-dae46d461369\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.923926 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-config\") pod \"43ea3e47-5f76-4668-84d7-dae46d461369\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.924098 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-dns-svc\") pod \"43ea3e47-5f76-4668-84d7-dae46d461369\" (UID: \"43ea3e47-5f76-4668-84d7-dae46d461369\") " Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.928404 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f2bac7de-ff6d-4b5f-85c0-c3653855f844-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.928437 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.928448 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n65w6\" (UniqueName: \"kubernetes.io/projected/f2bac7de-ff6d-4b5f-85c0-c3653855f844-kube-api-access-n65w6\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.928482 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thlj8\" (UniqueName: \"kubernetes.io/projected/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0-kube-api-access-thlj8\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.928493 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6n6r\" (UniqueName: \"kubernetes.io/projected/b24c1a1f-5324-4699-90a5-b539c855acb6-kube-api-access-g6n6r\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.928890 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43ea3e47-5f76-4668-84d7-dae46d461369-kube-api-access-bvc84" (OuterVolumeSpecName: "kube-api-access-bvc84") pod "43ea3e47-5f76-4668-84d7-dae46d461369" (UID: "43ea3e47-5f76-4668-84d7-dae46d461369"). InnerVolumeSpecName "kube-api-access-bvc84". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.929041 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-b544s" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.929187 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-b544s" event={"ID":"b24c1a1f-5324-4699-90a5-b539c855acb6","Type":"ContainerDied","Data":"234fdbdd6b1b2c1aaa1b61b22ff83450b84c90021fb7afc5d7b2fa0f33b09895"} Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.929241 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="234fdbdd6b1b2c1aaa1b61b22ff83450b84c90021fb7afc5d7b2fa0f33b09895" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.939200 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-b21d-account-create-cbzxv" event={"ID":"f2bac7de-ff6d-4b5f-85c0-c3653855f844","Type":"ContainerDied","Data":"85def9b7fd3bed2459efd73524d017099a937e7a39bd4ace2b1b5b1538e46d60"} Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.939250 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85def9b7fd3bed2459efd73524d017099a937e7a39bd4ace2b1b5b1538e46d60" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.939217 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-b21d-account-create-cbzxv" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.941344 4926 generic.go:334] "Generic (PLEG): container finished" podID="43ea3e47-5f76-4668-84d7-dae46d461369" containerID="7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0" exitCode=0 Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.941411 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.941414 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" event={"ID":"43ea3e47-5f76-4668-84d7-dae46d461369","Type":"ContainerDied","Data":"7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0"} Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.941519 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-fhgqq" event={"ID":"43ea3e47-5f76-4668-84d7-dae46d461369","Type":"ContainerDied","Data":"5a930120db87ed74e121c2fd0e9d988ae28ea4ea5cda0432e0961ac96a22abfa"} Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.941539 4926 scope.go:117] "RemoveContainer" containerID="7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.945187 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9vf4k" event={"ID":"c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0","Type":"ContainerDied","Data":"5d6f4602259a73d3250a9650db27d8b68b124fdfcf17abba5719cb066118fd7b"} Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.945226 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d6f4602259a73d3250a9650db27d8b68b124fdfcf17abba5719cb066118fd7b" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.945289 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9vf4k" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.973846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-config" (OuterVolumeSpecName: "config") pod "43ea3e47-5f76-4668-84d7-dae46d461369" (UID: "43ea3e47-5f76-4668-84d7-dae46d461369"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:21 crc kubenswrapper[4926]: I1125 10:05:21.993299 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43ea3e47-5f76-4668-84d7-dae46d461369" (UID: "43ea3e47-5f76-4668-84d7-dae46d461369"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.033399 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.033445 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvc84\" (UniqueName: \"kubernetes.io/projected/43ea3e47-5f76-4668-84d7-dae46d461369-kube-api-access-bvc84\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.033462 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43ea3e47-5f76-4668-84d7-dae46d461369-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.066358 4926 scope.go:117] "RemoveContainer" containerID="32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.105624 4926 scope.go:117] "RemoveContainer" containerID="7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0" Nov 25 10:05:22 crc kubenswrapper[4926]: E1125 10:05:22.106498 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0\": container with ID starting with 7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0 not found: ID does not exist" containerID="7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.106535 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0"} err="failed to get container status \"7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0\": rpc error: code = NotFound desc = could not find container \"7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0\": container with ID starting with 7c020a68f40fb34c5796459503eeaabba7a70bc1ab679a17a5baa2d84053d8f0 not found: ID does not exist" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.106559 4926 scope.go:117] "RemoveContainer" containerID="32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e" Nov 25 10:05:22 crc kubenswrapper[4926]: E1125 10:05:22.107064 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e\": container with ID starting with 32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e not found: ID does not exist" containerID="32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.107130 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e"} err="failed to get container status \"32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e\": rpc error: code = NotFound desc = could not find container \"32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e\": container with ID starting with 32f478c32ab0c51c80edf79cc43a36273b2602e2eeb04e074626cd74d539ed6e not found: ID does not exist" Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.265625 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhgqq"] Nov 25 10:05:22 crc kubenswrapper[4926]: I1125 10:05:22.269962 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-fhgqq"] Nov 25 10:05:24 crc kubenswrapper[4926]: I1125 10:05:24.024589 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" path="/var/lib/kubelet/pods/43ea3e47-5f76-4668-84d7-dae46d461369/volumes" Nov 25 10:05:29 crc kubenswrapper[4926]: I1125 10:05:29.030695 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6dkhn" podUID="af231b27-f9dd-466d-83d0-05484183099f" containerName="ovn-controller" probeResult="failure" output=< Nov 25 10:05:29 crc kubenswrapper[4926]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 10:05:29 crc kubenswrapper[4926]: > Nov 25 10:05:30 crc kubenswrapper[4926]: I1125 10:05:30.024876 4926 generic.go:334] "Generic (PLEG): container finished" podID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerID="838134005ec8486202b3ca3e5e28bed22a4b79db2906d9166c0e208a2fcef3c5" exitCode=0 Nov 25 10:05:30 crc kubenswrapper[4926]: I1125 10:05:30.025262 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78f39106-b4c8-41d0-b852-a12e64f5863b","Type":"ContainerDied","Data":"838134005ec8486202b3ca3e5e28bed22a4b79db2906d9166c0e208a2fcef3c5"} Nov 25 10:05:30 crc kubenswrapper[4926]: I1125 10:05:30.028391 4926 generic.go:334] "Generic (PLEG): container finished" podID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerID="07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b" exitCode=0 Nov 25 10:05:30 crc kubenswrapper[4926]: I1125 10:05:30.028422 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"64dbdcf7-3a3c-48c3-a68f-22644f250edd","Type":"ContainerDied","Data":"07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b"} Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.055974 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78f39106-b4c8-41d0-b852-a12e64f5863b","Type":"ContainerStarted","Data":"4a21748c09e0a6d16c38b18f9e819a289276f74dcd6dc74be0b5b0bac57419a5"} Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.056823 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.057987 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"64dbdcf7-3a3c-48c3-a68f-22644f250edd","Type":"ContainerStarted","Data":"caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380"} Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.058223 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.059554 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-grrv7" event={"ID":"ceab1d9b-89a4-412c-b720-b46193249d81","Type":"ContainerStarted","Data":"e6765602c9bbf756c103dde6d916c27b17acc611e585e6519f035172ff9e1063"} Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.088874 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=52.411522846 podStartE2EDuration="1m0.088850436s" podCreationTimestamp="2025-11-25 10:04:33 +0000 UTC" firstStartedPulling="2025-11-25 10:04:48.175637277 +0000 UTC m=+984.859593704" lastFinishedPulling="2025-11-25 10:04:55.852964867 +0000 UTC m=+992.536921294" observedRunningTime="2025-11-25 10:05:33.081400571 +0000 UTC m=+1029.765357038" watchObservedRunningTime="2025-11-25 10:05:33.088850436 +0000 UTC m=+1029.772806873" Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.114293 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=52.761953775 podStartE2EDuration="1m0.114259847s" podCreationTimestamp="2025-11-25 10:04:33 +0000 UTC" firstStartedPulling="2025-11-25 10:04:47.830348906 +0000 UTC m=+984.514305333" lastFinishedPulling="2025-11-25 10:04:55.182654978 +0000 UTC m=+991.866611405" observedRunningTime="2025-11-25 10:05:33.107311075 +0000 UTC m=+1029.791267512" watchObservedRunningTime="2025-11-25 10:05:33.114259847 +0000 UTC m=+1029.798216274" Nov 25 10:05:33 crc kubenswrapper[4926]: I1125 10:05:33.137408 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-grrv7" podStartSLOduration=2.838423424 podStartE2EDuration="15.137379061s" podCreationTimestamp="2025-11-25 10:05:18 +0000 UTC" firstStartedPulling="2025-11-25 10:05:19.560701586 +0000 UTC m=+1016.244658013" lastFinishedPulling="2025-11-25 10:05:31.859657223 +0000 UTC m=+1028.543613650" observedRunningTime="2025-11-25 10:05:33.131620508 +0000 UTC m=+1029.815576975" watchObservedRunningTime="2025-11-25 10:05:33.137379061 +0000 UTC m=+1029.821335498" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.023404 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-6dkhn" podUID="af231b27-f9dd-466d-83d0-05484183099f" containerName="ovn-controller" probeResult="failure" output=< Nov 25 10:05:34 crc kubenswrapper[4926]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Nov 25 10:05:34 crc kubenswrapper[4926]: > Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.030439 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.030717 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-dkcjx" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.270929 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-6dkhn-config-tpj69"] Nov 25 10:05:34 crc kubenswrapper[4926]: E1125 10:05:34.271417 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" containerName="dnsmasq-dns" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271443 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" containerName="dnsmasq-dns" Nov 25 10:05:34 crc kubenswrapper[4926]: E1125 10:05:34.271463 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b24c1a1f-5324-4699-90a5-b539c855acb6" containerName="mariadb-database-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271476 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b24c1a1f-5324-4699-90a5-b539c855acb6" containerName="mariadb-database-create" Nov 25 10:05:34 crc kubenswrapper[4926]: E1125 10:05:34.271501 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" containerName="mariadb-account-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271512 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" containerName="mariadb-account-create" Nov 25 10:05:34 crc kubenswrapper[4926]: E1125 10:05:34.271539 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" containerName="mariadb-database-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271549 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" containerName="mariadb-database-create" Nov 25 10:05:34 crc kubenswrapper[4926]: E1125 10:05:34.271570 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2bac7de-ff6d-4b5f-85c0-c3653855f844" containerName="mariadb-account-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271580 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2bac7de-ff6d-4b5f-85c0-c3653855f844" containerName="mariadb-account-create" Nov 25 10:05:34 crc kubenswrapper[4926]: E1125 10:05:34.271597 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" containerName="init" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271608 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" containerName="init" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271849 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" containerName="mariadb-account-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271872 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="43ea3e47-5f76-4668-84d7-dae46d461369" containerName="dnsmasq-dns" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271891 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2bac7de-ff6d-4b5f-85c0-c3653855f844" containerName="mariadb-account-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271915 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b24c1a1f-5324-4699-90a5-b539c855acb6" containerName="mariadb-database-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.271929 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" containerName="mariadb-database-create" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.272734 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.274979 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.289361 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6dkhn-config-tpj69"] Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.400068 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-log-ovn\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.400189 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t22n4\" (UniqueName: \"kubernetes.io/projected/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-kube-api-access-t22n4\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.400213 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run-ovn\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.400264 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.400326 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-additional-scripts\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.400365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-scripts\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.501525 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-log-ovn\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.501599 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t22n4\" (UniqueName: \"kubernetes.io/projected/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-kube-api-access-t22n4\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.501627 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run-ovn\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.501655 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.501689 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-additional-scripts\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.501736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-scripts\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.502257 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.502497 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-log-ovn\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.502899 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-additional-scripts\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.502182 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run-ovn\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.507314 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-scripts\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.525882 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t22n4\" (UniqueName: \"kubernetes.io/projected/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-kube-api-access-t22n4\") pod \"ovn-controller-6dkhn-config-tpj69\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:34 crc kubenswrapper[4926]: I1125 10:05:34.595220 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:35 crc kubenswrapper[4926]: I1125 10:05:35.108474 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-6dkhn-config-tpj69"] Nov 25 10:05:35 crc kubenswrapper[4926]: W1125 10:05:35.113123 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5ab8c09_cf4e_4a60_a6e7_4d5ee1e30e19.slice/crio-8d7074bffb0d1e4a244da31f658435b5a1cd888713fe4a49cf1135fdf30c09fa WatchSource:0}: Error finding container 8d7074bffb0d1e4a244da31f658435b5a1cd888713fe4a49cf1135fdf30c09fa: Status 404 returned error can't find the container with id 8d7074bffb0d1e4a244da31f658435b5a1cd888713fe4a49cf1135fdf30c09fa Nov 25 10:05:36 crc kubenswrapper[4926]: I1125 10:05:36.083276 4926 generic.go:334] "Generic (PLEG): container finished" podID="a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" containerID="07893fd6ab086c07011aedb868a2d874eb2ee0c6229d38efb7780171a6b1439b" exitCode=0 Nov 25 10:05:36 crc kubenswrapper[4926]: I1125 10:05:36.085030 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6dkhn-config-tpj69" event={"ID":"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19","Type":"ContainerDied","Data":"07893fd6ab086c07011aedb868a2d874eb2ee0c6229d38efb7780171a6b1439b"} Nov 25 10:05:36 crc kubenswrapper[4926]: I1125 10:05:36.085147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6dkhn-config-tpj69" event={"ID":"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19","Type":"ContainerStarted","Data":"8d7074bffb0d1e4a244da31f658435b5a1cd888713fe4a49cf1135fdf30c09fa"} Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.411431 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551144 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-additional-scripts\") pod \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551242 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-log-ovn\") pod \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551289 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t22n4\" (UniqueName: \"kubernetes.io/projected/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-kube-api-access-t22n4\") pod \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551349 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run\") pod \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551384 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" (UID: "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551395 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-scripts\") pod \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551453 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run-ovn\") pod \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\" (UID: \"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19\") " Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551481 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run" (OuterVolumeSpecName: "var-run") pod "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" (UID: "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.551595 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" (UID: "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.552267 4926 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-log-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.552301 4926 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.552313 4926 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-var-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.552435 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" (UID: "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.552703 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-scripts" (OuterVolumeSpecName: "scripts") pod "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" (UID: "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.561271 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-kube-api-access-t22n4" (OuterVolumeSpecName: "kube-api-access-t22n4") pod "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" (UID: "a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19"). InnerVolumeSpecName "kube-api-access-t22n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.653739 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t22n4\" (UniqueName: \"kubernetes.io/projected/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-kube-api-access-t22n4\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.654119 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:37 crc kubenswrapper[4926]: I1125 10:05:37.654133 4926 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19-additional-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.105861 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-6dkhn-config-tpj69" event={"ID":"a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19","Type":"ContainerDied","Data":"8d7074bffb0d1e4a244da31f658435b5a1cd888713fe4a49cf1135fdf30c09fa"} Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.105918 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d7074bffb0d1e4a244da31f658435b5a1cd888713fe4a49cf1135fdf30c09fa" Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.105912 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-6dkhn-config-tpj69" Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.109364 4926 generic.go:334] "Generic (PLEG): container finished" podID="ceab1d9b-89a4-412c-b720-b46193249d81" containerID="e6765602c9bbf756c103dde6d916c27b17acc611e585e6519f035172ff9e1063" exitCode=0 Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.109420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-grrv7" event={"ID":"ceab1d9b-89a4-412c-b720-b46193249d81","Type":"ContainerDied","Data":"e6765602c9bbf756c103dde6d916c27b17acc611e585e6519f035172ff9e1063"} Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.524168 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-6dkhn-config-tpj69"] Nov 25 10:05:38 crc kubenswrapper[4926]: I1125 10:05:38.529386 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-6dkhn-config-tpj69"] Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.001530 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-6dkhn" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.532407 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.686766 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-db-sync-config-data\") pod \"ceab1d9b-89a4-412c-b720-b46193249d81\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.687157 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpnh7\" (UniqueName: \"kubernetes.io/projected/ceab1d9b-89a4-412c-b720-b46193249d81-kube-api-access-fpnh7\") pod \"ceab1d9b-89a4-412c-b720-b46193249d81\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.687187 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-config-data\") pod \"ceab1d9b-89a4-412c-b720-b46193249d81\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.687369 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-combined-ca-bundle\") pod \"ceab1d9b-89a4-412c-b720-b46193249d81\" (UID: \"ceab1d9b-89a4-412c-b720-b46193249d81\") " Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.703183 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ceab1d9b-89a4-412c-b720-b46193249d81" (UID: "ceab1d9b-89a4-412c-b720-b46193249d81"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.708169 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ceab1d9b-89a4-412c-b720-b46193249d81-kube-api-access-fpnh7" (OuterVolumeSpecName: "kube-api-access-fpnh7") pod "ceab1d9b-89a4-412c-b720-b46193249d81" (UID: "ceab1d9b-89a4-412c-b720-b46193249d81"). InnerVolumeSpecName "kube-api-access-fpnh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.713615 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ceab1d9b-89a4-412c-b720-b46193249d81" (UID: "ceab1d9b-89a4-412c-b720-b46193249d81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.737048 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-config-data" (OuterVolumeSpecName: "config-data") pod "ceab1d9b-89a4-412c-b720-b46193249d81" (UID: "ceab1d9b-89a4-412c-b720-b46193249d81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.789093 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.789132 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.789142 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpnh7\" (UniqueName: \"kubernetes.io/projected/ceab1d9b-89a4-412c-b720-b46193249d81-kube-api-access-fpnh7\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:39 crc kubenswrapper[4926]: I1125 10:05:39.789153 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ceab1d9b-89a4-412c-b720-b46193249d81-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.028443 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" path="/var/lib/kubelet/pods/a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19/volumes" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.129176 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-grrv7" event={"ID":"ceab1d9b-89a4-412c-b720-b46193249d81","Type":"ContainerDied","Data":"2553640bff97827ea34a7e7913aa97034385ea5658dc747b75b56e6fad61ee36"} Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.129228 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2553640bff97827ea34a7e7913aa97034385ea5658dc747b75b56e6fad61ee36" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.129247 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-grrv7" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.673742 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-cx6sp"] Nov 25 10:05:40 crc kubenswrapper[4926]: E1125 10:05:40.675558 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ceab1d9b-89a4-412c-b720-b46193249d81" containerName="glance-db-sync" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.675643 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ceab1d9b-89a4-412c-b720-b46193249d81" containerName="glance-db-sync" Nov 25 10:05:40 crc kubenswrapper[4926]: E1125 10:05:40.675748 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" containerName="ovn-config" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.675838 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" containerName="ovn-config" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.676742 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ceab1d9b-89a4-412c-b720-b46193249d81" containerName="glance-db-sync" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.676805 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5ab8c09-cf4e-4a60-a6e7-4d5ee1e30e19" containerName="ovn-config" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.678543 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.711010 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-cx6sp"] Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.808215 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-nb\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.808316 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-dns-svc\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.808382 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-config\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.808409 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-sb\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.808431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvxlf\" (UniqueName: \"kubernetes.io/projected/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-kube-api-access-hvxlf\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.910073 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-dns-svc\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.910169 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-config\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.910199 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-sb\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.910217 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvxlf\" (UniqueName: \"kubernetes.io/projected/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-kube-api-access-hvxlf\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.910246 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-nb\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.911982 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-config\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.912459 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-sb\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.912787 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-dns-svc\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.912882 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-nb\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:40 crc kubenswrapper[4926]: I1125 10:05:40.933234 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvxlf\" (UniqueName: \"kubernetes.io/projected/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-kube-api-access-hvxlf\") pod \"dnsmasq-dns-554567b4f7-cx6sp\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:41 crc kubenswrapper[4926]: I1125 10:05:41.005544 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:41 crc kubenswrapper[4926]: I1125 10:05:41.545485 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-cx6sp"] Nov 25 10:05:42 crc kubenswrapper[4926]: I1125 10:05:42.154931 4926 generic.go:334] "Generic (PLEG): container finished" podID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerID="0a52d398fb7a37d6362da6d08c0a875106e4975c841917b0bbe11e423121b06b" exitCode=0 Nov 25 10:05:42 crc kubenswrapper[4926]: I1125 10:05:42.155047 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" event={"ID":"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd","Type":"ContainerDied","Data":"0a52d398fb7a37d6362da6d08c0a875106e4975c841917b0bbe11e423121b06b"} Nov 25 10:05:42 crc kubenswrapper[4926]: I1125 10:05:42.155289 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" event={"ID":"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd","Type":"ContainerStarted","Data":"748bbeb4b4548913947e846a1018482b45f0905878f914afc8ff5a96f21781f2"} Nov 25 10:05:43 crc kubenswrapper[4926]: I1125 10:05:43.167920 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" event={"ID":"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd","Type":"ContainerStarted","Data":"2386c47dc8b2f26c3e505f4a9d0159fbb3ac620d0d7346fa1f8a9387ab22c9db"} Nov 25 10:05:43 crc kubenswrapper[4926]: I1125 10:05:43.168148 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:43 crc kubenswrapper[4926]: I1125 10:05:43.193896 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" podStartSLOduration=3.193871741 podStartE2EDuration="3.193871741s" podCreationTimestamp="2025-11-25 10:05:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:05:43.187490153 +0000 UTC m=+1039.871446610" watchObservedRunningTime="2025-11-25 10:05:43.193871741 +0000 UTC m=+1039.877828178" Nov 25 10:05:44 crc kubenswrapper[4926]: I1125 10:05:44.637164 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:05:44 crc kubenswrapper[4926]: I1125 10:05:44.854388 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.070055 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-jvjxm"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.071286 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.082492 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jvjxm"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.178815 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-88zw6"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.183330 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.184437 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-d6aa-account-create-ksf77"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.185894 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.198598 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-operator-scripts\") pod \"cinder-db-create-jvjxm\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.198672 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rkwr\" (UniqueName: \"kubernetes.io/projected/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-kube-api-access-2rkwr\") pod \"cinder-db-create-jvjxm\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.199307 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.206665 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d6aa-account-create-ksf77"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.214355 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-88zw6"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.301007 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09f50a9c-beb0-4c48-9081-b740a6a0fc60-operator-scripts\") pod \"barbican-d6aa-account-create-ksf77\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.301074 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62jwx\" (UniqueName: \"kubernetes.io/projected/d9a07ca9-335b-484d-ba61-d9b4f798dd98-kube-api-access-62jwx\") pod \"barbican-db-create-88zw6\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.301143 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-operator-scripts\") pod \"cinder-db-create-jvjxm\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.301171 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a07ca9-335b-484d-ba61-d9b4f798dd98-operator-scripts\") pod \"barbican-db-create-88zw6\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.301206 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxmdp\" (UniqueName: \"kubernetes.io/projected/09f50a9c-beb0-4c48-9081-b740a6a0fc60-kube-api-access-lxmdp\") pod \"barbican-d6aa-account-create-ksf77\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.301229 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rkwr\" (UniqueName: \"kubernetes.io/projected/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-kube-api-access-2rkwr\") pod \"cinder-db-create-jvjxm\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.302026 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-operator-scripts\") pod \"cinder-db-create-jvjxm\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.309455 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-ea5b-account-create-4lqsw"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.310866 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.315734 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.344341 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea5b-account-create-4lqsw"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.346819 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rkwr\" (UniqueName: \"kubernetes.io/projected/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-kube-api-access-2rkwr\") pod \"cinder-db-create-jvjxm\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.391320 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.410272 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spdgb\" (UniqueName: \"kubernetes.io/projected/e1571b13-1388-4ce6-b85d-c3250984ea99-kube-api-access-spdgb\") pod \"cinder-ea5b-account-create-4lqsw\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.410381 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a07ca9-335b-484d-ba61-d9b4f798dd98-operator-scripts\") pod \"barbican-db-create-88zw6\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.410431 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxmdp\" (UniqueName: \"kubernetes.io/projected/09f50a9c-beb0-4c48-9081-b740a6a0fc60-kube-api-access-lxmdp\") pod \"barbican-d6aa-account-create-ksf77\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.410467 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1571b13-1388-4ce6-b85d-c3250984ea99-operator-scripts\") pod \"cinder-ea5b-account-create-4lqsw\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.410522 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09f50a9c-beb0-4c48-9081-b740a6a0fc60-operator-scripts\") pod \"barbican-d6aa-account-create-ksf77\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.410582 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62jwx\" (UniqueName: \"kubernetes.io/projected/d9a07ca9-335b-484d-ba61-d9b4f798dd98-kube-api-access-62jwx\") pod \"barbican-db-create-88zw6\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.411789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a07ca9-335b-484d-ba61-d9b4f798dd98-operator-scripts\") pod \"barbican-db-create-88zw6\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.412610 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09f50a9c-beb0-4c48-9081-b740a6a0fc60-operator-scripts\") pod \"barbican-d6aa-account-create-ksf77\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.417199 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-c66wr"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.424665 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.432057 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.432313 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.432461 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c77m5" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.432594 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.459144 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c66wr"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.469419 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxmdp\" (UniqueName: \"kubernetes.io/projected/09f50a9c-beb0-4c48-9081-b740a6a0fc60-kube-api-access-lxmdp\") pod \"barbican-d6aa-account-create-ksf77\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.481523 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62jwx\" (UniqueName: \"kubernetes.io/projected/d9a07ca9-335b-484d-ba61-d9b4f798dd98-kube-api-access-62jwx\") pod \"barbican-db-create-88zw6\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.505456 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.510712 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.512442 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-config-data\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.512494 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1571b13-1388-4ce6-b85d-c3250984ea99-operator-scripts\") pod \"cinder-ea5b-account-create-4lqsw\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.512546 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-428rj\" (UniqueName: \"kubernetes.io/projected/25d7aade-38a4-438b-a75d-420103b7bcdc-kube-api-access-428rj\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.512597 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spdgb\" (UniqueName: \"kubernetes.io/projected/e1571b13-1388-4ce6-b85d-c3250984ea99-kube-api-access-spdgb\") pod \"cinder-ea5b-account-create-4lqsw\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.512639 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-combined-ca-bundle\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.513355 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1571b13-1388-4ce6-b85d-c3250984ea99-operator-scripts\") pod \"cinder-ea5b-account-create-4lqsw\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.562641 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spdgb\" (UniqueName: \"kubernetes.io/projected/e1571b13-1388-4ce6-b85d-c3250984ea99-kube-api-access-spdgb\") pod \"cinder-ea5b-account-create-4lqsw\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.564030 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-nrx5l"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.565385 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.577061 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9d9f-account-create-n7kx2"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.578595 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.580599 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.584568 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nrx5l"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.603131 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9d9f-account-create-n7kx2"] Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.613893 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-combined-ca-bundle\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.613990 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-config-data\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.614059 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-428rj\" (UniqueName: \"kubernetes.io/projected/25d7aade-38a4-438b-a75d-420103b7bcdc-kube-api-access-428rj\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.622599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-combined-ca-bundle\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.624206 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-config-data\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.640264 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-428rj\" (UniqueName: \"kubernetes.io/projected/25d7aade-38a4-438b-a75d-420103b7bcdc-kube-api-access-428rj\") pod \"keystone-db-sync-c66wr\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.652151 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.717053 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-operator-scripts\") pod \"neutron-9d9f-account-create-n7kx2\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.717105 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6mk5\" (UniqueName: \"kubernetes.io/projected/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-kube-api-access-v6mk5\") pod \"neutron-9d9f-account-create-n7kx2\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.717174 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f66b61a5-3157-42ea-9dad-6b112c268960-operator-scripts\") pod \"neutron-db-create-nrx5l\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.717213 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfqm5\" (UniqueName: \"kubernetes.io/projected/f66b61a5-3157-42ea-9dad-6b112c268960-kube-api-access-zfqm5\") pod \"neutron-db-create-nrx5l\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.818565 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-operator-scripts\") pod \"neutron-9d9f-account-create-n7kx2\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.818651 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6mk5\" (UniqueName: \"kubernetes.io/projected/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-kube-api-access-v6mk5\") pod \"neutron-9d9f-account-create-n7kx2\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.818738 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f66b61a5-3157-42ea-9dad-6b112c268960-operator-scripts\") pod \"neutron-db-create-nrx5l\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.818780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfqm5\" (UniqueName: \"kubernetes.io/projected/f66b61a5-3157-42ea-9dad-6b112c268960-kube-api-access-zfqm5\") pod \"neutron-db-create-nrx5l\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.819680 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-operator-scripts\") pod \"neutron-9d9f-account-create-n7kx2\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.820278 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f66b61a5-3157-42ea-9dad-6b112c268960-operator-scripts\") pod \"neutron-db-create-nrx5l\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.843303 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfqm5\" (UniqueName: \"kubernetes.io/projected/f66b61a5-3157-42ea-9dad-6b112c268960-kube-api-access-zfqm5\") pod \"neutron-db-create-nrx5l\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.843860 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6mk5\" (UniqueName: \"kubernetes.io/projected/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-kube-api-access-v6mk5\") pod \"neutron-9d9f-account-create-n7kx2\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.892440 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.946878 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:45 crc kubenswrapper[4926]: I1125 10:05:45.977651 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.046679 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-ea5b-account-create-4lqsw"] Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.046723 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-d6aa-account-create-ksf77"] Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.070661 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-jvjxm"] Nov 25 10:05:46 crc kubenswrapper[4926]: W1125 10:05:46.082677 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecbca8cc_2f37_49d7_8038_a20c2d68e0eb.slice/crio-034973528c03b7069416513b34ee2eabd92c7adcaf07a14ef4613ec73851c53b WatchSource:0}: Error finding container 034973528c03b7069416513b34ee2eabd92c7adcaf07a14ef4613ec73851c53b: Status 404 returned error can't find the container with id 034973528c03b7069416513b34ee2eabd92c7adcaf07a14ef4613ec73851c53b Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.179668 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-88zw6"] Nov 25 10:05:46 crc kubenswrapper[4926]: W1125 10:05:46.206421 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9a07ca9_335b_484d_ba61_d9b4f798dd98.slice/crio-b0cd793d9ef499d2d536f1dc87334eab094c6f6369093b89ec0788f4db2c098c WatchSource:0}: Error finding container b0cd793d9ef499d2d536f1dc87334eab094c6f6369093b89ec0788f4db2c098c: Status 404 returned error can't find the container with id b0cd793d9ef499d2d536f1dc87334eab094c6f6369093b89ec0788f4db2c098c Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.215879 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea5b-account-create-4lqsw" event={"ID":"e1571b13-1388-4ce6-b85d-c3250984ea99","Type":"ContainerStarted","Data":"8b6225fff370b2b6f4a86590bd6d34452c16de79e0445be2f61fc9a5c4d664e1"} Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.248773 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvjxm" event={"ID":"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb","Type":"ContainerStarted","Data":"034973528c03b7069416513b34ee2eabd92c7adcaf07a14ef4613ec73851c53b"} Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.256700 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d6aa-account-create-ksf77" event={"ID":"09f50a9c-beb0-4c48-9081-b740a6a0fc60","Type":"ContainerStarted","Data":"2393b555859e972b703e1f436b5a4505b7f917ac9a7ed051b2c7526e5431bf1a"} Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.426107 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9d9f-account-create-n7kx2"] Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.513646 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-nrx5l"] Nov 25 10:05:46 crc kubenswrapper[4926]: I1125 10:05:46.537078 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-c66wr"] Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.271068 4926 generic.go:334] "Generic (PLEG): container finished" podID="09f50a9c-beb0-4c48-9081-b740a6a0fc60" containerID="b5415d070ca1c8776aa8d3d432c5e24a9220a9732426014a618911030b7c22cf" exitCode=0 Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.271154 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d6aa-account-create-ksf77" event={"ID":"09f50a9c-beb0-4c48-9081-b740a6a0fc60","Type":"ContainerDied","Data":"b5415d070ca1c8776aa8d3d432c5e24a9220a9732426014a618911030b7c22cf"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.273604 4926 generic.go:334] "Generic (PLEG): container finished" podID="e1571b13-1388-4ce6-b85d-c3250984ea99" containerID="b092152b6402577975beddfefb125dea0f49ebaf85b3d481068bfcad3e2a5709" exitCode=0 Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.273669 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea5b-account-create-4lqsw" event={"ID":"e1571b13-1388-4ce6-b85d-c3250984ea99","Type":"ContainerDied","Data":"b092152b6402577975beddfefb125dea0f49ebaf85b3d481068bfcad3e2a5709"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.278003 4926 generic.go:334] "Generic (PLEG): container finished" podID="ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" containerID="35ec8108122ed4ff7547d52a2e40d1e3ea39c8053a2374ba5c5e23b909a87970" exitCode=0 Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.278057 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvjxm" event={"ID":"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb","Type":"ContainerDied","Data":"35ec8108122ed4ff7547d52a2e40d1e3ea39c8053a2374ba5c5e23b909a87970"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.280239 4926 generic.go:334] "Generic (PLEG): container finished" podID="8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" containerID="3cccb390990e940eaa77f32f2e691b926b51ce574babcfe23b92f4df1aa76c28" exitCode=0 Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.280281 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9d9f-account-create-n7kx2" event={"ID":"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd","Type":"ContainerDied","Data":"3cccb390990e940eaa77f32f2e691b926b51ce574babcfe23b92f4df1aa76c28"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.280298 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9d9f-account-create-n7kx2" event={"ID":"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd","Type":"ContainerStarted","Data":"aedcb968f5326e582ccb0869941a92863130175e9eea3eb26fe6f80458285556"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.281758 4926 generic.go:334] "Generic (PLEG): container finished" podID="d9a07ca9-335b-484d-ba61-d9b4f798dd98" containerID="39f7fd2c0764b76e4a6cd1fef98fdcb90f7f6372ffcb67355393cef922e722bc" exitCode=0 Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.281803 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-88zw6" event={"ID":"d9a07ca9-335b-484d-ba61-d9b4f798dd98","Type":"ContainerDied","Data":"39f7fd2c0764b76e4a6cd1fef98fdcb90f7f6372ffcb67355393cef922e722bc"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.281823 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-88zw6" event={"ID":"d9a07ca9-335b-484d-ba61-d9b4f798dd98","Type":"ContainerStarted","Data":"b0cd793d9ef499d2d536f1dc87334eab094c6f6369093b89ec0788f4db2c098c"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.283413 4926 generic.go:334] "Generic (PLEG): container finished" podID="f66b61a5-3157-42ea-9dad-6b112c268960" containerID="6b2c2d5a9644a0e7be4f78d1877d23621dda6d82b88e85a3d71338fbf868a486" exitCode=0 Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.283468 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nrx5l" event={"ID":"f66b61a5-3157-42ea-9dad-6b112c268960","Type":"ContainerDied","Data":"6b2c2d5a9644a0e7be4f78d1877d23621dda6d82b88e85a3d71338fbf868a486"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.283487 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nrx5l" event={"ID":"f66b61a5-3157-42ea-9dad-6b112c268960","Type":"ContainerStarted","Data":"738cca52f4ee6b5aa79c09dbce181e84efed692657deec8b181ba8097b608c9e"} Nov 25 10:05:47 crc kubenswrapper[4926]: I1125 10:05:47.285807 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c66wr" event={"ID":"25d7aade-38a4-438b-a75d-420103b7bcdc","Type":"ContainerStarted","Data":"1d16c5d2b9b3b33b5c0d890a7e24b3e22f36ad90842eef5fcd7da0faa8bc15e3"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.007868 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.075544 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lfmd2"] Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.076767 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-lfmd2" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerName="dnsmasq-dns" containerID="cri-o://50538bb73cfc013a33e6a8e278d0a37af22444c3674886483ef937af8d9b7bb4" gracePeriod=10 Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.338831 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-nrx5l" event={"ID":"f66b61a5-3157-42ea-9dad-6b112c268960","Type":"ContainerDied","Data":"738cca52f4ee6b5aa79c09dbce181e84efed692657deec8b181ba8097b608c9e"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.338877 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="738cca52f4ee6b5aa79c09dbce181e84efed692657deec8b181ba8097b608c9e" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.340512 4926 generic.go:334] "Generic (PLEG): container finished" podID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerID="50538bb73cfc013a33e6a8e278d0a37af22444c3674886483ef937af8d9b7bb4" exitCode=0 Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.340566 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lfmd2" event={"ID":"9b82a532-0d0e-4cf3-892c-540122644a0e","Type":"ContainerDied","Data":"50538bb73cfc013a33e6a8e278d0a37af22444c3674886483ef937af8d9b7bb4"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.347966 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-d6aa-account-create-ksf77" event={"ID":"09f50a9c-beb0-4c48-9081-b740a6a0fc60","Type":"ContainerDied","Data":"2393b555859e972b703e1f436b5a4505b7f917ac9a7ed051b2c7526e5431bf1a"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.348033 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2393b555859e972b703e1f436b5a4505b7f917ac9a7ed051b2c7526e5431bf1a" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.349698 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-ea5b-account-create-4lqsw" event={"ID":"e1571b13-1388-4ce6-b85d-c3250984ea99","Type":"ContainerDied","Data":"8b6225fff370b2b6f4a86590bd6d34452c16de79e0445be2f61fc9a5c4d664e1"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.349723 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b6225fff370b2b6f4a86590bd6d34452c16de79e0445be2f61fc9a5c4d664e1" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.352371 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-jvjxm" event={"ID":"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb","Type":"ContainerDied","Data":"034973528c03b7069416513b34ee2eabd92c7adcaf07a14ef4613ec73851c53b"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.352455 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="034973528c03b7069416513b34ee2eabd92c7adcaf07a14ef4613ec73851c53b" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.358581 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9d9f-account-create-n7kx2" event={"ID":"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd","Type":"ContainerDied","Data":"aedcb968f5326e582ccb0869941a92863130175e9eea3eb26fe6f80458285556"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.358645 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aedcb968f5326e582ccb0869941a92863130175e9eea3eb26fe6f80458285556" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.361998 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-88zw6" event={"ID":"d9a07ca9-335b-484d-ba61-d9b4f798dd98","Type":"ContainerDied","Data":"b0cd793d9ef499d2d536f1dc87334eab094c6f6369093b89ec0788f4db2c098c"} Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.362050 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0cd793d9ef499d2d536f1dc87334eab094c6f6369093b89ec0788f4db2c098c" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.362876 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.394553 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.408229 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.428109 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.433627 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.452694 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rkwr\" (UniqueName: \"kubernetes.io/projected/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-kube-api-access-2rkwr\") pod \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.452795 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-operator-scripts\") pod \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\" (UID: \"ecbca8cc-2f37-49d7-8038-a20c2d68e0eb\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.453572 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" (UID: "ecbca8cc-2f37-49d7-8038-a20c2d68e0eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.456186 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.459747 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-kube-api-access-2rkwr" (OuterVolumeSpecName: "kube-api-access-2rkwr") pod "ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" (UID: "ecbca8cc-2f37-49d7-8038-a20c2d68e0eb"). InnerVolumeSpecName "kube-api-access-2rkwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554204 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1571b13-1388-4ce6-b85d-c3250984ea99-operator-scripts\") pod \"e1571b13-1388-4ce6-b85d-c3250984ea99\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554263 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f66b61a5-3157-42ea-9dad-6b112c268960-operator-scripts\") pod \"f66b61a5-3157-42ea-9dad-6b112c268960\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554327 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-operator-scripts\") pod \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554405 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spdgb\" (UniqueName: \"kubernetes.io/projected/e1571b13-1388-4ce6-b85d-c3250984ea99-kube-api-access-spdgb\") pod \"e1571b13-1388-4ce6-b85d-c3250984ea99\" (UID: \"e1571b13-1388-4ce6-b85d-c3250984ea99\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554429 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6mk5\" (UniqueName: \"kubernetes.io/projected/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-kube-api-access-v6mk5\") pod \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\" (UID: \"8a65fcf0-3c1c-4832-a039-ddd082b5c7bd\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554457 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62jwx\" (UniqueName: \"kubernetes.io/projected/d9a07ca9-335b-484d-ba61-d9b4f798dd98-kube-api-access-62jwx\") pod \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554527 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfqm5\" (UniqueName: \"kubernetes.io/projected/f66b61a5-3157-42ea-9dad-6b112c268960-kube-api-access-zfqm5\") pod \"f66b61a5-3157-42ea-9dad-6b112c268960\" (UID: \"f66b61a5-3157-42ea-9dad-6b112c268960\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554557 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxmdp\" (UniqueName: \"kubernetes.io/projected/09f50a9c-beb0-4c48-9081-b740a6a0fc60-kube-api-access-lxmdp\") pod \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554617 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a07ca9-335b-484d-ba61-d9b4f798dd98-operator-scripts\") pod \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\" (UID: \"d9a07ca9-335b-484d-ba61-d9b4f798dd98\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554643 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09f50a9c-beb0-4c48-9081-b740a6a0fc60-operator-scripts\") pod \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\" (UID: \"09f50a9c-beb0-4c48-9081-b740a6a0fc60\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554928 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rkwr\" (UniqueName: \"kubernetes.io/projected/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-kube-api-access-2rkwr\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.554964 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.555401 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09f50a9c-beb0-4c48-9081-b740a6a0fc60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09f50a9c-beb0-4c48-9081-b740a6a0fc60" (UID: "09f50a9c-beb0-4c48-9081-b740a6a0fc60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.555732 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1571b13-1388-4ce6-b85d-c3250984ea99-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e1571b13-1388-4ce6-b85d-c3250984ea99" (UID: "e1571b13-1388-4ce6-b85d-c3250984ea99"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.556110 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f66b61a5-3157-42ea-9dad-6b112c268960-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f66b61a5-3157-42ea-9dad-6b112c268960" (UID: "f66b61a5-3157-42ea-9dad-6b112c268960"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.557049 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" (UID: "8a65fcf0-3c1c-4832-a039-ddd082b5c7bd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.558200 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9a07ca9-335b-484d-ba61-d9b4f798dd98-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d9a07ca9-335b-484d-ba61-d9b4f798dd98" (UID: "d9a07ca9-335b-484d-ba61-d9b4f798dd98"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.559869 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9a07ca9-335b-484d-ba61-d9b4f798dd98-kube-api-access-62jwx" (OuterVolumeSpecName: "kube-api-access-62jwx") pod "d9a07ca9-335b-484d-ba61-d9b4f798dd98" (UID: "d9a07ca9-335b-484d-ba61-d9b4f798dd98"). InnerVolumeSpecName "kube-api-access-62jwx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.561274 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-kube-api-access-v6mk5" (OuterVolumeSpecName: "kube-api-access-v6mk5") pod "8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" (UID: "8a65fcf0-3c1c-4832-a039-ddd082b5c7bd"). InnerVolumeSpecName "kube-api-access-v6mk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.561320 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09f50a9c-beb0-4c48-9081-b740a6a0fc60-kube-api-access-lxmdp" (OuterVolumeSpecName: "kube-api-access-lxmdp") pod "09f50a9c-beb0-4c48-9081-b740a6a0fc60" (UID: "09f50a9c-beb0-4c48-9081-b740a6a0fc60"). InnerVolumeSpecName "kube-api-access-lxmdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.561613 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.562652 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1571b13-1388-4ce6-b85d-c3250984ea99-kube-api-access-spdgb" (OuterVolumeSpecName: "kube-api-access-spdgb") pod "e1571b13-1388-4ce6-b85d-c3250984ea99" (UID: "e1571b13-1388-4ce6-b85d-c3250984ea99"). InnerVolumeSpecName "kube-api-access-spdgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.565638 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f66b61a5-3157-42ea-9dad-6b112c268960-kube-api-access-zfqm5" (OuterVolumeSpecName: "kube-api-access-zfqm5") pod "f66b61a5-3157-42ea-9dad-6b112c268960" (UID: "f66b61a5-3157-42ea-9dad-6b112c268960"). InnerVolumeSpecName "kube-api-access-zfqm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.656149 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-sb\") pod \"9b82a532-0d0e-4cf3-892c-540122644a0e\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.656262 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-dns-svc\") pod \"9b82a532-0d0e-4cf3-892c-540122644a0e\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.656386 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-config\") pod \"9b82a532-0d0e-4cf3-892c-540122644a0e\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.656413 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqk68\" (UniqueName: \"kubernetes.io/projected/9b82a532-0d0e-4cf3-892c-540122644a0e-kube-api-access-hqk68\") pod \"9b82a532-0d0e-4cf3-892c-540122644a0e\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.656459 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-nb\") pod \"9b82a532-0d0e-4cf3-892c-540122644a0e\" (UID: \"9b82a532-0d0e-4cf3-892c-540122644a0e\") " Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.656980 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657001 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spdgb\" (UniqueName: \"kubernetes.io/projected/e1571b13-1388-4ce6-b85d-c3250984ea99-kube-api-access-spdgb\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657016 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6mk5\" (UniqueName: \"kubernetes.io/projected/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd-kube-api-access-v6mk5\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657056 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62jwx\" (UniqueName: \"kubernetes.io/projected/d9a07ca9-335b-484d-ba61-d9b4f798dd98-kube-api-access-62jwx\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657069 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfqm5\" (UniqueName: \"kubernetes.io/projected/f66b61a5-3157-42ea-9dad-6b112c268960-kube-api-access-zfqm5\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657082 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxmdp\" (UniqueName: \"kubernetes.io/projected/09f50a9c-beb0-4c48-9081-b740a6a0fc60-kube-api-access-lxmdp\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657094 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d9a07ca9-335b-484d-ba61-d9b4f798dd98-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657131 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09f50a9c-beb0-4c48-9081-b740a6a0fc60-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657143 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e1571b13-1388-4ce6-b85d-c3250984ea99-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.657155 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f66b61a5-3157-42ea-9dad-6b112c268960-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.661057 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b82a532-0d0e-4cf3-892c-540122644a0e-kube-api-access-hqk68" (OuterVolumeSpecName: "kube-api-access-hqk68") pod "9b82a532-0d0e-4cf3-892c-540122644a0e" (UID: "9b82a532-0d0e-4cf3-892c-540122644a0e"). InnerVolumeSpecName "kube-api-access-hqk68". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.702813 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9b82a532-0d0e-4cf3-892c-540122644a0e" (UID: "9b82a532-0d0e-4cf3-892c-540122644a0e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.704318 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9b82a532-0d0e-4cf3-892c-540122644a0e" (UID: "9b82a532-0d0e-4cf3-892c-540122644a0e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.712695 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9b82a532-0d0e-4cf3-892c-540122644a0e" (UID: "9b82a532-0d0e-4cf3-892c-540122644a0e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.715748 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-config" (OuterVolumeSpecName: "config") pod "9b82a532-0d0e-4cf3-892c-540122644a0e" (UID: "9b82a532-0d0e-4cf3-892c-540122644a0e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.758598 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.758632 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.758642 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.758652 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqk68\" (UniqueName: \"kubernetes.io/projected/9b82a532-0d0e-4cf3-892c-540122644a0e-kube-api-access-hqk68\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:51 crc kubenswrapper[4926]: I1125 10:05:51.758664 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9b82a532-0d0e-4cf3-892c-540122644a0e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.373131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c66wr" event={"ID":"25d7aade-38a4-438b-a75d-420103b7bcdc","Type":"ContainerStarted","Data":"2503491034c10378141b596e162960a67b1158b9e06d0a02b70f3a18377cb69d"} Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.376969 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-lfmd2" event={"ID":"9b82a532-0d0e-4cf3-892c-540122644a0e","Type":"ContainerDied","Data":"cc9df3c5be74e3566c485f35398453098bb02a196f2a9ecbe343ca47cbc9b9ce"} Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377045 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9d9f-account-create-n7kx2" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377059 4926 scope.go:117] "RemoveContainer" containerID="50538bb73cfc013a33e6a8e278d0a37af22444c3674886483ef937af8d9b7bb4" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377130 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-nrx5l" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377130 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-jvjxm" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377214 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-d6aa-account-create-ksf77" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377353 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-ea5b-account-create-4lqsw" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.377834 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-88zw6" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.378061 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-lfmd2" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.429805 4926 scope.go:117] "RemoveContainer" containerID="3a896daa993688f15b9f07a8138bf33cdd5f905689a38a5d83bf483079a3ce27" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.442836 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-c66wr" podStartSLOduration=2.872423731 podStartE2EDuration="7.442816488s" podCreationTimestamp="2025-11-25 10:05:45 +0000 UTC" firstStartedPulling="2025-11-25 10:05:46.574192074 +0000 UTC m=+1043.258148501" lastFinishedPulling="2025-11-25 10:05:51.144584831 +0000 UTC m=+1047.828541258" observedRunningTime="2025-11-25 10:05:52.40705672 +0000 UTC m=+1049.091013157" watchObservedRunningTime="2025-11-25 10:05:52.442816488 +0000 UTC m=+1049.126772905" Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.490196 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lfmd2"] Nov 25 10:05:52 crc kubenswrapper[4926]: I1125 10:05:52.502044 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-lfmd2"] Nov 25 10:05:54 crc kubenswrapper[4926]: I1125 10:05:54.028517 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" path="/var/lib/kubelet/pods/9b82a532-0d0e-4cf3-892c-540122644a0e/volumes" Nov 25 10:05:54 crc kubenswrapper[4926]: I1125 10:05:54.403191 4926 generic.go:334] "Generic (PLEG): container finished" podID="25d7aade-38a4-438b-a75d-420103b7bcdc" containerID="2503491034c10378141b596e162960a67b1158b9e06d0a02b70f3a18377cb69d" exitCode=0 Nov 25 10:05:54 crc kubenswrapper[4926]: I1125 10:05:54.403232 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c66wr" event={"ID":"25d7aade-38a4-438b-a75d-420103b7bcdc","Type":"ContainerDied","Data":"2503491034c10378141b596e162960a67b1158b9e06d0a02b70f3a18377cb69d"} Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.761275 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.834096 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-combined-ca-bundle\") pod \"25d7aade-38a4-438b-a75d-420103b7bcdc\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.834189 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-config-data\") pod \"25d7aade-38a4-438b-a75d-420103b7bcdc\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.834357 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-428rj\" (UniqueName: \"kubernetes.io/projected/25d7aade-38a4-438b-a75d-420103b7bcdc-kube-api-access-428rj\") pod \"25d7aade-38a4-438b-a75d-420103b7bcdc\" (UID: \"25d7aade-38a4-438b-a75d-420103b7bcdc\") " Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.841290 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25d7aade-38a4-438b-a75d-420103b7bcdc-kube-api-access-428rj" (OuterVolumeSpecName: "kube-api-access-428rj") pod "25d7aade-38a4-438b-a75d-420103b7bcdc" (UID: "25d7aade-38a4-438b-a75d-420103b7bcdc"). InnerVolumeSpecName "kube-api-access-428rj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.857931 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "25d7aade-38a4-438b-a75d-420103b7bcdc" (UID: "25d7aade-38a4-438b-a75d-420103b7bcdc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.879873 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-config-data" (OuterVolumeSpecName: "config-data") pod "25d7aade-38a4-438b-a75d-420103b7bcdc" (UID: "25d7aade-38a4-438b-a75d-420103b7bcdc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.936806 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.936871 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/25d7aade-38a4-438b-a75d-420103b7bcdc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:55 crc kubenswrapper[4926]: I1125 10:05:55.936931 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-428rj\" (UniqueName: \"kubernetes.io/projected/25d7aade-38a4-438b-a75d-420103b7bcdc-kube-api-access-428rj\") on node \"crc\" DevicePath \"\"" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.427587 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-c66wr" event={"ID":"25d7aade-38a4-438b-a75d-420103b7bcdc","Type":"ContainerDied","Data":"1d16c5d2b9b3b33b5c0d890a7e24b3e22f36ad90842eef5fcd7da0faa8bc15e3"} Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.427868 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d16c5d2b9b3b33b5c0d890a7e24b3e22f36ad90842eef5fcd7da0faa8bc15e3" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.427667 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-c66wr" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.702734 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xx6j8"] Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703182 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerName="init" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703201 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerName="init" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703219 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25d7aade-38a4-438b-a75d-420103b7bcdc" containerName="keystone-db-sync" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703227 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="25d7aade-38a4-438b-a75d-420103b7bcdc" containerName="keystone-db-sync" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703243 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703254 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703273 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerName="dnsmasq-dns" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703282 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerName="dnsmasq-dns" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703312 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f66b61a5-3157-42ea-9dad-6b112c268960" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703323 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f66b61a5-3157-42ea-9dad-6b112c268960" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703338 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9a07ca9-335b-484d-ba61-d9b4f798dd98" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703347 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9a07ca9-335b-484d-ba61-d9b4f798dd98" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703360 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09f50a9c-beb0-4c48-9081-b740a6a0fc60" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703368 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="09f50a9c-beb0-4c48-9081-b740a6a0fc60" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703383 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703393 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: E1125 10:05:56.703405 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1571b13-1388-4ce6-b85d-c3250984ea99" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703413 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1571b13-1388-4ce6-b85d-c3250984ea99" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703604 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1571b13-1388-4ce6-b85d-c3250984ea99" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703615 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703629 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b82a532-0d0e-4cf3-892c-540122644a0e" containerName="dnsmasq-dns" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703641 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f66b61a5-3157-42ea-9dad-6b112c268960" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703651 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9a07ca9-335b-484d-ba61-d9b4f798dd98" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703666 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="09f50a9c-beb0-4c48-9081-b740a6a0fc60" containerName="mariadb-account-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703681 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" containerName="mariadb-database-create" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.703693 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="25d7aade-38a4-438b-a75d-420103b7bcdc" containerName="keystone-db-sync" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.704863 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.721948 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-8qxdn"] Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.722915 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.729455 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c77m5" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.729788 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.730184 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.730745 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.735188 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.752016 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-sb\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.752083 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-config\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.752111 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xscd8\" (UniqueName: \"kubernetes.io/projected/8ba28479-336e-4364-92c4-c63fb1e54e2c-kube-api-access-xscd8\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.752522 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-dns-svc\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.752676 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.769243 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8qxdn"] Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.790738 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xx6j8"] Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854220 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-combined-ca-bundle\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854287 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-sb\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854318 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-credential-keys\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854350 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-config\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854418 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-config-data\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854482 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xscd8\" (UniqueName: \"kubernetes.io/projected/8ba28479-336e-4364-92c4-c63fb1e54e2c-kube-api-access-xscd8\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854527 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mf59k\" (UniqueName: \"kubernetes.io/projected/dfe6f9df-d193-4de7-877f-5899e2e8df78-kube-api-access-mf59k\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854638 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-fernet-keys\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-scripts\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-dns-svc\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.854782 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.855365 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-config\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.855689 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-sb\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.855744 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-nb\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.856495 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-dns-svc\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.880885 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xscd8\" (UniqueName: \"kubernetes.io/projected/8ba28479-336e-4364-92c4-c63fb1e54e2c-kube-api-access-xscd8\") pod \"dnsmasq-dns-67795cd9-xx6j8\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.956233 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-combined-ca-bundle\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.956282 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-credential-keys\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.956305 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-config-data\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.956337 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mf59k\" (UniqueName: \"kubernetes.io/projected/dfe6f9df-d193-4de7-877f-5899e2e8df78-kube-api-access-mf59k\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.956379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-fernet-keys\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.956408 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-scripts\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.957750 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-k6rtr"] Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.959366 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.960309 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-scripts\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.960911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-fernet-keys\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.961379 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-config-data\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.968462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-credential-keys\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.975856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-combined-ca-bundle\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.978183 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.978676 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.978841 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nzjdg" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.982590 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mf59k\" (UniqueName: \"kubernetes.io/projected/dfe6f9df-d193-4de7-877f-5899e2e8df78-kube-api-access-mf59k\") pod \"keystone-bootstrap-8qxdn\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:56 crc kubenswrapper[4926]: I1125 10:05:56.986989 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-k6rtr"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.020075 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-lslvn"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.021164 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.023101 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.029445 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.029502 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.029676 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5jv27" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.043125 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lslvn"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.044050 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.052060 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-cjmn2"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.059922 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-config\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.059988 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-combined-ca-bundle\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060009 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-db-sync-config-data\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060026 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-combined-ca-bundle\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060069 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6c095-3a19-4536-b55a-c79c991badd0-etc-machine-id\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060085 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwtsj\" (UniqueName: \"kubernetes.io/projected/65e6c095-3a19-4536-b55a-c79c991badd0-kube-api-access-hwtsj\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060107 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-scripts\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060143 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl5x4\" (UniqueName: \"kubernetes.io/projected/eb63e765-910d-4764-942c-bec556b98a5e-kube-api-access-rl5x4\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.060172 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-config-data\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.061447 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.065598 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.065822 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wxp4z" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.073427 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.079145 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cjmn2"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.100488 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.103389 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.107316 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.116058 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.148770 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xx6j8"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.163831 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-log-httpd\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.163894 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-config-data\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.163912 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfvrg\" (UniqueName: \"kubernetes.io/projected/170174f9-d802-42f5-97a1-13c6bbe50460-kube-api-access-zfvrg\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.163976 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-config\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.163994 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-config-data\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164018 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73e4f7-21d9-4d05-b900-1e49498b4874-logs\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164034 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-scripts\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164063 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-combined-ca-bundle\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164083 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-db-sync-config-data\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164109 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-combined-ca-bundle\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164131 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-combined-ca-bundle\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164164 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6c095-3a19-4536-b55a-c79c991badd0-etc-machine-id\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwtsj\" (UniqueName: \"kubernetes.io/projected/65e6c095-3a19-4536-b55a-c79c991badd0-kube-api-access-hwtsj\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164205 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-scripts\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164223 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164242 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-run-httpd\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164264 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164293 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-scripts\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164312 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl5x4\" (UniqueName: \"kubernetes.io/projected/eb63e765-910d-4764-942c-bec556b98a5e-kube-api-access-rl5x4\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164329 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-config-data\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.164347 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx2v9\" (UniqueName: \"kubernetes.io/projected/8c73e4f7-21d9-4d05-b900-1e49498b4874-kube-api-access-dx2v9\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.168552 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6c095-3a19-4536-b55a-c79c991badd0-etc-machine-id\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.172750 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.173094 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-db-sync-config-data\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.175032 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-k58gt"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.176462 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-config\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.176692 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.179556 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-combined-ca-bundle\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.180348 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-config-data\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.180883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-combined-ca-bundle\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.185543 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fc6gp" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.185558 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.193879 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl5x4\" (UniqueName: \"kubernetes.io/projected/eb63e765-910d-4764-942c-bec556b98a5e-kube-api-access-rl5x4\") pod \"neutron-db-sync-lslvn\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.195640 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-scripts\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.200716 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-k58gt"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.207576 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwtsj\" (UniqueName: \"kubernetes.io/projected/65e6c095-3a19-4536-b55a-c79c991badd0-kube-api-access-hwtsj\") pod \"cinder-db-sync-k6rtr\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.236988 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.238441 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.268862 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270072 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270129 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-scripts\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270179 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-combined-ca-bundle\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270203 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-config-data\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270232 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx2v9\" (UniqueName: \"kubernetes.io/projected/8c73e4f7-21d9-4d05-b900-1e49498b4874-kube-api-access-dx2v9\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270260 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-log-httpd\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270286 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfvrg\" (UniqueName: \"kubernetes.io/projected/170174f9-d802-42f5-97a1-13c6bbe50460-kube-api-access-zfvrg\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270317 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9k9p7\" (UniqueName: \"kubernetes.io/projected/b4001327-fff9-4fe5-9ae8-3a47dea56a57-kube-api-access-9k9p7\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270340 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270383 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-config-data\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270411 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270438 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9crcl\" (UniqueName: \"kubernetes.io/projected/af3d71a6-8937-4d95-9152-d77c1afc39eb-kube-api-access-9crcl\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270463 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73e4f7-21d9-4d05-b900-1e49498b4874-logs\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270486 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-scripts\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270514 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-db-sync-config-data\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270575 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270607 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-combined-ca-bundle\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270663 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270686 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-config\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.270710 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-run-httpd\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.271343 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-run-httpd\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.273786 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-log-httpd\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.274361 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73e4f7-21d9-4d05-b900-1e49498b4874-logs\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.277402 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-combined-ca-bundle\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.278192 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-config-data\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.278248 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-scripts\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.278307 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-config-data\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.283672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.283799 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-scripts\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.292802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.298622 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx2v9\" (UniqueName: \"kubernetes.io/projected/8c73e4f7-21d9-4d05-b900-1e49498b4874-kube-api-access-dx2v9\") pod \"placement-db-sync-cjmn2\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.300431 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfvrg\" (UniqueName: \"kubernetes.io/projected/170174f9-d802-42f5-97a1-13c6bbe50460-kube-api-access-zfvrg\") pod \"ceilometer-0\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.359540 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.373929 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-combined-ca-bundle\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374047 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9k9p7\" (UniqueName: \"kubernetes.io/projected/b4001327-fff9-4fe5-9ae8-3a47dea56a57-kube-api-access-9k9p7\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374073 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374105 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374128 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9crcl\" (UniqueName: \"kubernetes.io/projected/af3d71a6-8937-4d95-9152-d77c1afc39eb-kube-api-access-9crcl\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374157 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-db-sync-config-data\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374194 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.374248 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-config\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.377558 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-config\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.383789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-combined-ca-bundle\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.384108 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-sb\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.384452 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-dns-svc\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.385502 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-nb\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.387390 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-db-sync-config-data\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.413311 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9k9p7\" (UniqueName: \"kubernetes.io/projected/b4001327-fff9-4fe5-9ae8-3a47dea56a57-kube-api-access-9k9p7\") pod \"dnsmasq-dns-5b6dbdb6f5-k9ln5\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.413570 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9crcl\" (UniqueName: \"kubernetes.io/projected/af3d71a6-8937-4d95-9152-d77c1afc39eb-kube-api-access-9crcl\") pod \"barbican-db-sync-k58gt\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.448381 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lslvn" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.469084 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cjmn2" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.493477 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.501271 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k58gt" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.584351 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.740504 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xx6j8"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.768286 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-8qxdn"] Nov 25 10:05:57 crc kubenswrapper[4926]: I1125 10:05:57.795925 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-k6rtr"] Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.206811 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-k58gt"] Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.226039 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-cjmn2"] Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.252009 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5"] Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.265324 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lslvn"] Nov 25 10:05:58 crc kubenswrapper[4926]: W1125 10:05:58.428561 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod170174f9_d802_42f5_97a1_13c6bbe50460.slice/crio-07c526368b5f6cfe3724a2306cf45413dc7ab120df963117a94928fec723df46 WatchSource:0}: Error finding container 07c526368b5f6cfe3724a2306cf45413dc7ab120df963117a94928fec723df46: Status 404 returned error can't find the container with id 07c526368b5f6cfe3724a2306cf45413dc7ab120df963117a94928fec723df46 Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.431821 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.459372 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" event={"ID":"8ba28479-336e-4364-92c4-c63fb1e54e2c","Type":"ContainerStarted","Data":"79af810595435d51f370668583ca3f632b0254e1e4c02e8f3650cfbf6e61bba7"} Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.460739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" event={"ID":"b4001327-fff9-4fe5-9ae8-3a47dea56a57","Type":"ContainerStarted","Data":"ccf39292d0fde2c6797e1daf7a324d06b320aec91e9ecbcb507a99fa7235d54d"} Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.476729 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lslvn" event={"ID":"eb63e765-910d-4764-942c-bec556b98a5e","Type":"ContainerStarted","Data":"09da2d2d0c5e752c6ca6017b787e9c13194e6b5cbc79631bea64e07be4ac5622"} Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.479272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cjmn2" event={"ID":"8c73e4f7-21d9-4d05-b900-1e49498b4874","Type":"ContainerStarted","Data":"4f5bd394243c8d6b67bc0f6b35506f2a0d2a0bb3d59b77e19762a9bfd4234b88"} Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.484726 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8qxdn" event={"ID":"dfe6f9df-d193-4de7-877f-5899e2e8df78","Type":"ContainerStarted","Data":"230eee79dd2a4b279bd04749190437df1b2be5bd1140d7d03bee07bd70744864"} Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.486467 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k58gt" event={"ID":"af3d71a6-8937-4d95-9152-d77c1afc39eb","Type":"ContainerStarted","Data":"5fc08b2d5d6de8c4d0abe67f4df52ed55bf3adbd442321c150b3ae68b50cf125"} Nov 25 10:05:58 crc kubenswrapper[4926]: I1125 10:05:58.487262 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k6rtr" event={"ID":"65e6c095-3a19-4536-b55a-c79c991badd0","Type":"ContainerStarted","Data":"2f21aa166b9c7c23a51d8f906b99dce0e46c696c0b14bb1fe88bc97640ae40ec"} Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.364704 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.519806 4926 generic.go:334] "Generic (PLEG): container finished" podID="8ba28479-336e-4364-92c4-c63fb1e54e2c" containerID="38a1f501302b7d99fd6095aaf3d296ef080ae4d27abac8448f8402c138a3cccd" exitCode=0 Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.519890 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" event={"ID":"8ba28479-336e-4364-92c4-c63fb1e54e2c","Type":"ContainerDied","Data":"38a1f501302b7d99fd6095aaf3d296ef080ae4d27abac8448f8402c138a3cccd"} Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.527117 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerStarted","Data":"07c526368b5f6cfe3724a2306cf45413dc7ab120df963117a94928fec723df46"} Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.531327 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lslvn" event={"ID":"eb63e765-910d-4764-942c-bec556b98a5e","Type":"ContainerStarted","Data":"e9533031e180f6e5b0ae28a161aec469b550ffce26995a2f7cdae035505f1001"} Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.534268 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerID="f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559" exitCode=0 Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.534316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" event={"ID":"b4001327-fff9-4fe5-9ae8-3a47dea56a57","Type":"ContainerDied","Data":"f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559"} Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.587431 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8qxdn" event={"ID":"dfe6f9df-d193-4de7-877f-5899e2e8df78","Type":"ContainerStarted","Data":"7ba6c86f234e2a08f3839ab902b2e09e8268a1369b066b66413dd1fbc6f64bbb"} Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.610356 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-lslvn" podStartSLOduration=3.610293472 podStartE2EDuration="3.610293472s" podCreationTimestamp="2025-11-25 10:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:05:59.589205409 +0000 UTC m=+1056.273161836" watchObservedRunningTime="2025-11-25 10:05:59.610293472 +0000 UTC m=+1056.294249899" Nov 25 10:05:59 crc kubenswrapper[4926]: I1125 10:05:59.685880 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-8qxdn" podStartSLOduration=3.685858768 podStartE2EDuration="3.685858768s" podCreationTimestamp="2025-11-25 10:05:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:05:59.633625922 +0000 UTC m=+1056.317582349" watchObservedRunningTime="2025-11-25 10:05:59.685858768 +0000 UTC m=+1056.369815195" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.066973 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.139442 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-nb\") pod \"8ba28479-336e-4364-92c4-c63fb1e54e2c\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.139528 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xscd8\" (UniqueName: \"kubernetes.io/projected/8ba28479-336e-4364-92c4-c63fb1e54e2c-kube-api-access-xscd8\") pod \"8ba28479-336e-4364-92c4-c63fb1e54e2c\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.139618 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-config\") pod \"8ba28479-336e-4364-92c4-c63fb1e54e2c\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.139648 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-dns-svc\") pod \"8ba28479-336e-4364-92c4-c63fb1e54e2c\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.139767 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-sb\") pod \"8ba28479-336e-4364-92c4-c63fb1e54e2c\" (UID: \"8ba28479-336e-4364-92c4-c63fb1e54e2c\") " Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.162176 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ba28479-336e-4364-92c4-c63fb1e54e2c-kube-api-access-xscd8" (OuterVolumeSpecName: "kube-api-access-xscd8") pod "8ba28479-336e-4364-92c4-c63fb1e54e2c" (UID: "8ba28479-336e-4364-92c4-c63fb1e54e2c"). InnerVolumeSpecName "kube-api-access-xscd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.178762 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8ba28479-336e-4364-92c4-c63fb1e54e2c" (UID: "8ba28479-336e-4364-92c4-c63fb1e54e2c"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.181399 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-config" (OuterVolumeSpecName: "config") pod "8ba28479-336e-4364-92c4-c63fb1e54e2c" (UID: "8ba28479-336e-4364-92c4-c63fb1e54e2c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.238963 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8ba28479-336e-4364-92c4-c63fb1e54e2c" (UID: "8ba28479-336e-4364-92c4-c63fb1e54e2c"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.243751 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.243784 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.243794 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.243804 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xscd8\" (UniqueName: \"kubernetes.io/projected/8ba28479-336e-4364-92c4-c63fb1e54e2c-kube-api-access-xscd8\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.262623 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8ba28479-336e-4364-92c4-c63fb1e54e2c" (UID: "8ba28479-336e-4364-92c4-c63fb1e54e2c"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.345615 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8ba28479-336e-4364-92c4-c63fb1e54e2c-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.600168 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" event={"ID":"b4001327-fff9-4fe5-9ae8-3a47dea56a57","Type":"ContainerStarted","Data":"1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c"} Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.600238 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.603601 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.604228 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67795cd9-xx6j8" event={"ID":"8ba28479-336e-4364-92c4-c63fb1e54e2c","Type":"ContainerDied","Data":"79af810595435d51f370668583ca3f632b0254e1e4c02e8f3650cfbf6e61bba7"} Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.604309 4926 scope.go:117] "RemoveContainer" containerID="38a1f501302b7d99fd6095aaf3d296ef080ae4d27abac8448f8402c138a3cccd" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.623758 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" podStartSLOduration=3.623743109 podStartE2EDuration="3.623743109s" podCreationTimestamp="2025-11-25 10:05:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:00.620107489 +0000 UTC m=+1057.304063916" watchObservedRunningTime="2025-11-25 10:06:00.623743109 +0000 UTC m=+1057.307699536" Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.676051 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xx6j8"] Nov 25 10:06:00 crc kubenswrapper[4926]: I1125 10:06:00.682432 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67795cd9-xx6j8"] Nov 25 10:06:02 crc kubenswrapper[4926]: I1125 10:06:02.025710 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ba28479-336e-4364-92c4-c63fb1e54e2c" path="/var/lib/kubelet/pods/8ba28479-336e-4364-92c4-c63fb1e54e2c/volumes" Nov 25 10:06:02 crc kubenswrapper[4926]: I1125 10:06:02.635138 4926 generic.go:334] "Generic (PLEG): container finished" podID="dfe6f9df-d193-4de7-877f-5899e2e8df78" containerID="7ba6c86f234e2a08f3839ab902b2e09e8268a1369b066b66413dd1fbc6f64bbb" exitCode=0 Nov 25 10:06:02 crc kubenswrapper[4926]: I1125 10:06:02.635192 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8qxdn" event={"ID":"dfe6f9df-d193-4de7-877f-5899e2e8df78","Type":"ContainerDied","Data":"7ba6c86f234e2a08f3839ab902b2e09e8268a1369b066b66413dd1fbc6f64bbb"} Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.399738 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.490205 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-combined-ca-bundle\") pod \"dfe6f9df-d193-4de7-877f-5899e2e8df78\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.490269 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-scripts\") pod \"dfe6f9df-d193-4de7-877f-5899e2e8df78\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.490403 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mf59k\" (UniqueName: \"kubernetes.io/projected/dfe6f9df-d193-4de7-877f-5899e2e8df78-kube-api-access-mf59k\") pod \"dfe6f9df-d193-4de7-877f-5899e2e8df78\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.490476 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-credential-keys\") pod \"dfe6f9df-d193-4de7-877f-5899e2e8df78\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.490666 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-fernet-keys\") pod \"dfe6f9df-d193-4de7-877f-5899e2e8df78\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.490717 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-config-data\") pod \"dfe6f9df-d193-4de7-877f-5899e2e8df78\" (UID: \"dfe6f9df-d193-4de7-877f-5899e2e8df78\") " Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.497062 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfe6f9df-d193-4de7-877f-5899e2e8df78-kube-api-access-mf59k" (OuterVolumeSpecName: "kube-api-access-mf59k") pod "dfe6f9df-d193-4de7-877f-5899e2e8df78" (UID: "dfe6f9df-d193-4de7-877f-5899e2e8df78"). InnerVolumeSpecName "kube-api-access-mf59k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.497121 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-scripts" (OuterVolumeSpecName: "scripts") pod "dfe6f9df-d193-4de7-877f-5899e2e8df78" (UID: "dfe6f9df-d193-4de7-877f-5899e2e8df78"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.501065 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "dfe6f9df-d193-4de7-877f-5899e2e8df78" (UID: "dfe6f9df-d193-4de7-877f-5899e2e8df78"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.501274 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dfe6f9df-d193-4de7-877f-5899e2e8df78" (UID: "dfe6f9df-d193-4de7-877f-5899e2e8df78"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.520559 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-config-data" (OuterVolumeSpecName: "config-data") pod "dfe6f9df-d193-4de7-877f-5899e2e8df78" (UID: "dfe6f9df-d193-4de7-877f-5899e2e8df78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.522759 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dfe6f9df-d193-4de7-877f-5899e2e8df78" (UID: "dfe6f9df-d193-4de7-877f-5899e2e8df78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.586072 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.592564 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.592628 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.592641 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mf59k\" (UniqueName: \"kubernetes.io/projected/dfe6f9df-d193-4de7-877f-5899e2e8df78-kube-api-access-mf59k\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.592653 4926 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.592663 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.592674 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dfe6f9df-d193-4de7-877f-5899e2e8df78-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.653771 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-cx6sp"] Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.654078 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="dnsmasq-dns" containerID="cri-o://2386c47dc8b2f26c3e505f4a9d0159fbb3ac620d0d7346fa1f8a9387ab22c9db" gracePeriod=10 Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.679314 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-8qxdn" event={"ID":"dfe6f9df-d193-4de7-877f-5899e2e8df78","Type":"ContainerDied","Data":"230eee79dd2a4b279bd04749190437df1b2be5bd1140d7d03bee07bd70744864"} Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.679357 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="230eee79dd2a4b279bd04749190437df1b2be5bd1140d7d03bee07bd70744864" Nov 25 10:06:07 crc kubenswrapper[4926]: I1125 10:06:07.679415 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-8qxdn" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.478799 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-8qxdn"] Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.484883 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-8qxdn"] Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.590447 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-9fjls"] Nov 25 10:06:08 crc kubenswrapper[4926]: E1125 10:06:08.590847 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ba28479-336e-4364-92c4-c63fb1e54e2c" containerName="init" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.590867 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ba28479-336e-4364-92c4-c63fb1e54e2c" containerName="init" Nov 25 10:06:08 crc kubenswrapper[4926]: E1125 10:06:08.590888 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfe6f9df-d193-4de7-877f-5899e2e8df78" containerName="keystone-bootstrap" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.590897 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfe6f9df-d193-4de7-877f-5899e2e8df78" containerName="keystone-bootstrap" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.591133 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfe6f9df-d193-4de7-877f-5899e2e8df78" containerName="keystone-bootstrap" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.591153 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ba28479-336e-4364-92c4-c63fb1e54e2c" containerName="init" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.592116 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.595876 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.596225 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.596422 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.596572 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.596713 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c77m5" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.607281 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9fjls"] Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.688248 4926 generic.go:334] "Generic (PLEG): container finished" podID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerID="2386c47dc8b2f26c3e505f4a9d0159fbb3ac620d0d7346fa1f8a9387ab22c9db" exitCode=0 Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.688293 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" event={"ID":"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd","Type":"ContainerDied","Data":"2386c47dc8b2f26c3e505f4a9d0159fbb3ac620d0d7346fa1f8a9387ab22c9db"} Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.715501 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-combined-ca-bundle\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.715648 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qx92\" (UniqueName: \"kubernetes.io/projected/545481d3-3362-45f8-929b-71e8c75490dc-kube-api-access-5qx92\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.715693 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-scripts\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.715726 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-fernet-keys\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.715753 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-credential-keys\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.716046 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-config-data\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.817616 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-combined-ca-bundle\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.817723 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-scripts\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.817746 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qx92\" (UniqueName: \"kubernetes.io/projected/545481d3-3362-45f8-929b-71e8c75490dc-kube-api-access-5qx92\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.817772 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-fernet-keys\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.817792 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-credential-keys\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.817830 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-config-data\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.829862 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-scripts\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.830327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-credential-keys\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.830369 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-config-data\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.830460 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-fernet-keys\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.838843 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-combined-ca-bundle\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.840223 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qx92\" (UniqueName: \"kubernetes.io/projected/545481d3-3362-45f8-929b-71e8c75490dc-kube-api-access-5qx92\") pod \"keystone-bootstrap-9fjls\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:08 crc kubenswrapper[4926]: I1125 10:06:08.921744 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:10 crc kubenswrapper[4926]: I1125 10:06:10.024289 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfe6f9df-d193-4de7-877f-5899e2e8df78" path="/var/lib/kubelet/pods/dfe6f9df-d193-4de7-877f-5899e2e8df78/volumes" Nov 25 10:06:14 crc kubenswrapper[4926]: I1125 10:06:14.019176 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:06:14 crc kubenswrapper[4926]: I1125 10:06:14.019228 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:06:16 crc kubenswrapper[4926]: I1125 10:06:16.006741 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Nov 25 10:06:20 crc kubenswrapper[4926]: E1125 10:06:20.298162 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Nov 25 10:06:20 crc kubenswrapper[4926]: E1125 10:06:20.298856 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9crcl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-k58gt_openstack(af3d71a6-8937-4d95-9152-d77c1afc39eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:06:20 crc kubenswrapper[4926]: E1125 10:06:20.300011 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-k58gt" podUID="af3d71a6-8937-4d95-9152-d77c1afc39eb" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.456112 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.538343 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-dns-svc\") pod \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.538407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-config\") pod \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.538472 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-nb\") pod \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.538530 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvxlf\" (UniqueName: \"kubernetes.io/projected/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-kube-api-access-hvxlf\") pod \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.538561 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-sb\") pod \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\" (UID: \"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd\") " Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.543857 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-kube-api-access-hvxlf" (OuterVolumeSpecName: "kube-api-access-hvxlf") pod "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" (UID: "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd"). InnerVolumeSpecName "kube-api-access-hvxlf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.579906 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" (UID: "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.581833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" (UID: "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.597092 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" (UID: "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.602003 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-config" (OuterVolumeSpecName: "config") pod "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" (UID: "48fd2bd3-3e71-4dfb-8dab-dfb403da1afd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.640926 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.640978 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.640993 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.641007 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvxlf\" (UniqueName: \"kubernetes.io/projected/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-kube-api-access-hvxlf\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.641019 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.795903 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" event={"ID":"48fd2bd3-3e71-4dfb-8dab-dfb403da1afd","Type":"ContainerDied","Data":"748bbeb4b4548913947e846a1018482b45f0905878f914afc8ff5a96f21781f2"} Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.795967 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.795994 4926 scope.go:117] "RemoveContainer" containerID="2386c47dc8b2f26c3e505f4a9d0159fbb3ac620d0d7346fa1f8a9387ab22c9db" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.798752 4926 generic.go:334] "Generic (PLEG): container finished" podID="eb63e765-910d-4764-942c-bec556b98a5e" containerID="e9533031e180f6e5b0ae28a161aec469b550ffce26995a2f7cdae035505f1001" exitCode=0 Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.798772 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lslvn" event={"ID":"eb63e765-910d-4764-942c-bec556b98a5e","Type":"ContainerDied","Data":"e9533031e180f6e5b0ae28a161aec469b550ffce26995a2f7cdae035505f1001"} Nov 25 10:06:20 crc kubenswrapper[4926]: E1125 10:06:20.800848 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-k58gt" podUID="af3d71a6-8937-4d95-9152-d77c1afc39eb" Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.855724 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-cx6sp"] Nov 25 10:06:20 crc kubenswrapper[4926]: I1125 10:06:20.861315 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-554567b4f7-cx6sp"] Nov 25 10:06:21 crc kubenswrapper[4926]: I1125 10:06:21.010107 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-554567b4f7-cx6sp" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Nov 25 10:06:21 crc kubenswrapper[4926]: I1125 10:06:21.407393 4926 scope.go:117] "RemoveContainer" containerID="0a52d398fb7a37d6362da6d08c0a875106e4975c841917b0bbe11e423121b06b" Nov 25 10:06:21 crc kubenswrapper[4926]: E1125 10:06:21.456174 4926 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Nov 25 10:06:21 crc kubenswrapper[4926]: E1125 10:06:21.456449 4926 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwtsj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-k6rtr_openstack(65e6c095-3a19-4536-b55a-c79c991badd0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 25 10:06:21 crc kubenswrapper[4926]: E1125 10:06:21.457682 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-k6rtr" podUID="65e6c095-3a19-4536-b55a-c79c991badd0" Nov 25 10:06:21 crc kubenswrapper[4926]: I1125 10:06:21.809381 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerStarted","Data":"5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5"} Nov 25 10:06:21 crc kubenswrapper[4926]: I1125 10:06:21.811055 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cjmn2" event={"ID":"8c73e4f7-21d9-4d05-b900-1e49498b4874","Type":"ContainerStarted","Data":"d9288af6d592a262ca6d7f702f7a6c4c0d258a53e7b71269db47ea91ecc56728"} Nov 25 10:06:21 crc kubenswrapper[4926]: E1125 10:06:21.813648 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-k6rtr" podUID="65e6c095-3a19-4536-b55a-c79c991badd0" Nov 25 10:06:21 crc kubenswrapper[4926]: I1125 10:06:21.824988 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-9fjls"] Nov 25 10:06:21 crc kubenswrapper[4926]: I1125 10:06:21.838802 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-cjmn2" podStartSLOduration=1.6985942600000001 podStartE2EDuration="24.838776617s" podCreationTimestamp="2025-11-25 10:05:57 +0000 UTC" firstStartedPulling="2025-11-25 10:05:58.248120528 +0000 UTC m=+1054.932076955" lastFinishedPulling="2025-11-25 10:06:21.388302885 +0000 UTC m=+1078.072259312" observedRunningTime="2025-11-25 10:06:21.828578494 +0000 UTC m=+1078.512534921" watchObservedRunningTime="2025-11-25 10:06:21.838776617 +0000 UTC m=+1078.522733044" Nov 25 10:06:21 crc kubenswrapper[4926]: W1125 10:06:21.841885 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod545481d3_3362_45f8_929b_71e8c75490dc.slice/crio-c01dfdf3f2e0a28206c87655a20c64a5631263c8d046784ae77ad10bfc7251ab WatchSource:0}: Error finding container c01dfdf3f2e0a28206c87655a20c64a5631263c8d046784ae77ad10bfc7251ab: Status 404 returned error can't find the container with id c01dfdf3f2e0a28206c87655a20c64a5631263c8d046784ae77ad10bfc7251ab Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.023850 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" path="/var/lib/kubelet/pods/48fd2bd3-3e71-4dfb-8dab-dfb403da1afd/volumes" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.063978 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lslvn" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.167509 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-combined-ca-bundle\") pod \"eb63e765-910d-4764-942c-bec556b98a5e\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.167902 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rl5x4\" (UniqueName: \"kubernetes.io/projected/eb63e765-910d-4764-942c-bec556b98a5e-kube-api-access-rl5x4\") pod \"eb63e765-910d-4764-942c-bec556b98a5e\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.168137 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-config\") pod \"eb63e765-910d-4764-942c-bec556b98a5e\" (UID: \"eb63e765-910d-4764-942c-bec556b98a5e\") " Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.171457 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb63e765-910d-4764-942c-bec556b98a5e-kube-api-access-rl5x4" (OuterVolumeSpecName: "kube-api-access-rl5x4") pod "eb63e765-910d-4764-942c-bec556b98a5e" (UID: "eb63e765-910d-4764-942c-bec556b98a5e"). InnerVolumeSpecName "kube-api-access-rl5x4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.188132 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb63e765-910d-4764-942c-bec556b98a5e" (UID: "eb63e765-910d-4764-942c-bec556b98a5e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.189520 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-config" (OuterVolumeSpecName: "config") pod "eb63e765-910d-4764-942c-bec556b98a5e" (UID: "eb63e765-910d-4764-942c-bec556b98a5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.270696 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.270758 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rl5x4\" (UniqueName: \"kubernetes.io/projected/eb63e765-910d-4764-942c-bec556b98a5e-kube-api-access-rl5x4\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.270788 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/eb63e765-910d-4764-942c-bec556b98a5e-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.849577 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9fjls" event={"ID":"545481d3-3362-45f8-929b-71e8c75490dc","Type":"ContainerStarted","Data":"8f1ac120062adb0fbd34758fc7c784ee691a7138d7e707fcb64648157b587bd0"} Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.850160 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9fjls" event={"ID":"545481d3-3362-45f8-929b-71e8c75490dc","Type":"ContainerStarted","Data":"c01dfdf3f2e0a28206c87655a20c64a5631263c8d046784ae77ad10bfc7251ab"} Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.858279 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lslvn" event={"ID":"eb63e765-910d-4764-942c-bec556b98a5e","Type":"ContainerDied","Data":"09da2d2d0c5e752c6ca6017b787e9c13194e6b5cbc79631bea64e07be4ac5622"} Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.858342 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09da2d2d0c5e752c6ca6017b787e9c13194e6b5cbc79631bea64e07be4ac5622" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.858459 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lslvn" Nov 25 10:06:22 crc kubenswrapper[4926]: I1125 10:06:22.871875 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-9fjls" podStartSLOduration=14.871855702 podStartE2EDuration="14.871855702s" podCreationTimestamp="2025-11-25 10:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:22.86934948 +0000 UTC m=+1079.553305907" watchObservedRunningTime="2025-11-25 10:06:22.871855702 +0000 UTC m=+1079.555812129" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.051867 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-r7r2g"] Nov 25 10:06:23 crc kubenswrapper[4926]: E1125 10:06:23.052567 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb63e765-910d-4764-942c-bec556b98a5e" containerName="neutron-db-sync" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.052587 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb63e765-910d-4764-942c-bec556b98a5e" containerName="neutron-db-sync" Nov 25 10:06:23 crc kubenswrapper[4926]: E1125 10:06:23.052608 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="init" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.052616 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="init" Nov 25 10:06:23 crc kubenswrapper[4926]: E1125 10:06:23.052638 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="dnsmasq-dns" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.052648 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="dnsmasq-dns" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.052957 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb63e765-910d-4764-942c-bec556b98a5e" containerName="neutron-db-sync" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.052984 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="48fd2bd3-3e71-4dfb-8dab-dfb403da1afd" containerName="dnsmasq-dns" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.055097 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.078859 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-r7r2g"] Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.085198 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.085251 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vstw\" (UniqueName: \"kubernetes.io/projected/14ff3b9a-634a-433c-a454-3d09f6fddbf0-kube-api-access-8vstw\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.085311 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.085376 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-config\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.085460 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.171344 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76b9fdf556-hhq5f"] Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.172967 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.176088 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.176088 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5jv27" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.176172 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.181635 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76b9fdf556-hhq5f"] Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.184483 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.186887 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-config\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.186959 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-combined-ca-bundle\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187019 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-config\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187069 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-ovndb-tls-certs\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187180 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187207 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-httpd-config\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187238 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vstw\" (UniqueName: \"kubernetes.io/projected/14ff3b9a-634a-433c-a454-3d09f6fddbf0-kube-api-access-8vstw\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffq7d\" (UniqueName: \"kubernetes.io/projected/408f4021-7f18-46a6-b6e8-ef300c2f176a-kube-api-access-ffq7d\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.187332 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.188014 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-sb\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.188061 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-config\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.188185 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-nb\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.188748 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-dns-svc\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.227838 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vstw\" (UniqueName: \"kubernetes.io/projected/14ff3b9a-634a-433c-a454-3d09f6fddbf0-kube-api-access-8vstw\") pod \"dnsmasq-dns-5f66db59b9-r7r2g\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.288254 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-combined-ca-bundle\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.288323 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-config\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.288380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-ovndb-tls-certs\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.288404 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-httpd-config\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.288430 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffq7d\" (UniqueName: \"kubernetes.io/projected/408f4021-7f18-46a6-b6e8-ef300c2f176a-kube-api-access-ffq7d\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.293447 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-ovndb-tls-certs\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.294298 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-config\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.294817 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-httpd-config\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.295102 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-combined-ca-bundle\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.309228 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffq7d\" (UniqueName: \"kubernetes.io/projected/408f4021-7f18-46a6-b6e8-ef300c2f176a-kube-api-access-ffq7d\") pod \"neutron-76b9fdf556-hhq5f\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.372115 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.511422 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:23 crc kubenswrapper[4926]: W1125 10:06:23.831057 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod14ff3b9a_634a_433c_a454_3d09f6fddbf0.slice/crio-8cd32010ad7a8cc16d6b59f4b89ace02aac3a03b2b70b39987378df5fdef26dd WatchSource:0}: Error finding container 8cd32010ad7a8cc16d6b59f4b89ace02aac3a03b2b70b39987378df5fdef26dd: Status 404 returned error can't find the container with id 8cd32010ad7a8cc16d6b59f4b89ace02aac3a03b2b70b39987378df5fdef26dd Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.832042 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-r7r2g"] Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.870553 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" event={"ID":"14ff3b9a-634a-433c-a454-3d09f6fddbf0","Type":"ContainerStarted","Data":"8cd32010ad7a8cc16d6b59f4b89ace02aac3a03b2b70b39987378df5fdef26dd"} Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.872731 4926 generic.go:334] "Generic (PLEG): container finished" podID="8c73e4f7-21d9-4d05-b900-1e49498b4874" containerID="d9288af6d592a262ca6d7f702f7a6c4c0d258a53e7b71269db47ea91ecc56728" exitCode=0 Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.872788 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cjmn2" event={"ID":"8c73e4f7-21d9-4d05-b900-1e49498b4874","Type":"ContainerDied","Data":"d9288af6d592a262ca6d7f702f7a6c4c0d258a53e7b71269db47ea91ecc56728"} Nov 25 10:06:23 crc kubenswrapper[4926]: I1125 10:06:23.877474 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerStarted","Data":"e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442"} Nov 25 10:06:24 crc kubenswrapper[4926]: W1125 10:06:24.098193 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod408f4021_7f18_46a6_b6e8_ef300c2f176a.slice/crio-1665be9afb3ead929a0e6c15650895bd2f659807b3001fc797c875f258dd28ac WatchSource:0}: Error finding container 1665be9afb3ead929a0e6c15650895bd2f659807b3001fc797c875f258dd28ac: Status 404 returned error can't find the container with id 1665be9afb3ead929a0e6c15650895bd2f659807b3001fc797c875f258dd28ac Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.098877 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76b9fdf556-hhq5f"] Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.891929 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b9fdf556-hhq5f" event={"ID":"408f4021-7f18-46a6-b6e8-ef300c2f176a","Type":"ContainerStarted","Data":"b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a"} Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.892283 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b9fdf556-hhq5f" event={"ID":"408f4021-7f18-46a6-b6e8-ef300c2f176a","Type":"ContainerStarted","Data":"03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0"} Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.892298 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b9fdf556-hhq5f" event={"ID":"408f4021-7f18-46a6-b6e8-ef300c2f176a","Type":"ContainerStarted","Data":"1665be9afb3ead929a0e6c15650895bd2f659807b3001fc797c875f258dd28ac"} Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.893588 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.898750 4926 generic.go:334] "Generic (PLEG): container finished" podID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerID="7aa1f95a630fa036b347fb8ffd5d3f1b67a018db89a6b91ecb06548144f58364" exitCode=0 Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.899422 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" event={"ID":"14ff3b9a-634a-433c-a454-3d09f6fddbf0","Type":"ContainerDied","Data":"7aa1f95a630fa036b347fb8ffd5d3f1b67a018db89a6b91ecb06548144f58364"} Nov 25 10:06:24 crc kubenswrapper[4926]: I1125 10:06:24.933932 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76b9fdf556-hhq5f" podStartSLOduration=1.93391562 podStartE2EDuration="1.93391562s" podCreationTimestamp="2025-11-25 10:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:24.930018293 +0000 UTC m=+1081.613974740" watchObservedRunningTime="2025-11-25 10:06:24.93391562 +0000 UTC m=+1081.617872047" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.273482 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cjmn2" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.449765 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73e4f7-21d9-4d05-b900-1e49498b4874-logs\") pod \"8c73e4f7-21d9-4d05-b900-1e49498b4874\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.449819 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-scripts\") pod \"8c73e4f7-21d9-4d05-b900-1e49498b4874\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.449916 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-combined-ca-bundle\") pod \"8c73e4f7-21d9-4d05-b900-1e49498b4874\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.450502 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c73e4f7-21d9-4d05-b900-1e49498b4874-logs" (OuterVolumeSpecName: "logs") pod "8c73e4f7-21d9-4d05-b900-1e49498b4874" (UID: "8c73e4f7-21d9-4d05-b900-1e49498b4874"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.451168 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-config-data\") pod \"8c73e4f7-21d9-4d05-b900-1e49498b4874\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.451327 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dx2v9\" (UniqueName: \"kubernetes.io/projected/8c73e4f7-21d9-4d05-b900-1e49498b4874-kube-api-access-dx2v9\") pod \"8c73e4f7-21d9-4d05-b900-1e49498b4874\" (UID: \"8c73e4f7-21d9-4d05-b900-1e49498b4874\") " Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.452508 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8c73e4f7-21d9-4d05-b900-1e49498b4874-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.459246 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-scripts" (OuterVolumeSpecName: "scripts") pod "8c73e4f7-21d9-4d05-b900-1e49498b4874" (UID: "8c73e4f7-21d9-4d05-b900-1e49498b4874"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.471754 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c73e4f7-21d9-4d05-b900-1e49498b4874-kube-api-access-dx2v9" (OuterVolumeSpecName: "kube-api-access-dx2v9") pod "8c73e4f7-21d9-4d05-b900-1e49498b4874" (UID: "8c73e4f7-21d9-4d05-b900-1e49498b4874"). InnerVolumeSpecName "kube-api-access-dx2v9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.485427 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8c73e4f7-21d9-4d05-b900-1e49498b4874" (UID: "8c73e4f7-21d9-4d05-b900-1e49498b4874"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.513505 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-config-data" (OuterVolumeSpecName: "config-data") pod "8c73e4f7-21d9-4d05-b900-1e49498b4874" (UID: "8c73e4f7-21d9-4d05-b900-1e49498b4874"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.550998 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6cdc678f6f-htchg"] Nov 25 10:06:25 crc kubenswrapper[4926]: E1125 10:06:25.551513 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c73e4f7-21d9-4d05-b900-1e49498b4874" containerName="placement-db-sync" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.551541 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c73e4f7-21d9-4d05-b900-1e49498b4874" containerName="placement-db-sync" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.551750 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c73e4f7-21d9-4d05-b900-1e49498b4874" containerName="placement-db-sync" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.553025 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.554790 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.555031 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.559731 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-ovndb-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.559965 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-combined-ca-bundle\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560048 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-httpd-config\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560097 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95trz\" (UniqueName: \"kubernetes.io/projected/331a3347-d9a9-4a86-a552-58ab450406d4-kube-api-access-95trz\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560130 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-public-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560282 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-internal-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560329 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-config\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560852 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560911 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560926 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dx2v9\" (UniqueName: \"kubernetes.io/projected/8c73e4f7-21d9-4d05-b900-1e49498b4874-kube-api-access-dx2v9\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.560983 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8c73e4f7-21d9-4d05-b900-1e49498b4874-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.563909 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cdc678f6f-htchg"] Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662470 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-internal-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662530 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-config\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662661 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-ovndb-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662737 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-combined-ca-bundle\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662776 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-httpd-config\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662807 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95trz\" (UniqueName: \"kubernetes.io/projected/331a3347-d9a9-4a86-a552-58ab450406d4-kube-api-access-95trz\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.662830 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-public-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.668394 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-public-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.670431 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-internal-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.673215 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-combined-ca-bundle\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.673571 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-ovndb-tls-certs\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.675028 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-config\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.675957 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/331a3347-d9a9-4a86-a552-58ab450406d4-httpd-config\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.690676 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95trz\" (UniqueName: \"kubernetes.io/projected/331a3347-d9a9-4a86-a552-58ab450406d4-kube-api-access-95trz\") pod \"neutron-6cdc678f6f-htchg\" (UID: \"331a3347-d9a9-4a86-a552-58ab450406d4\") " pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.916966 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.918757 4926 generic.go:334] "Generic (PLEG): container finished" podID="545481d3-3362-45f8-929b-71e8c75490dc" containerID="8f1ac120062adb0fbd34758fc7c784ee691a7138d7e707fcb64648157b587bd0" exitCode=0 Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.918790 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9fjls" event={"ID":"545481d3-3362-45f8-929b-71e8c75490dc","Type":"ContainerDied","Data":"8f1ac120062adb0fbd34758fc7c784ee691a7138d7e707fcb64648157b587bd0"} Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.923719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-cjmn2" event={"ID":"8c73e4f7-21d9-4d05-b900-1e49498b4874","Type":"ContainerDied","Data":"4f5bd394243c8d6b67bc0f6b35506f2a0d2a0bb3d59b77e19762a9bfd4234b88"} Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.923783 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f5bd394243c8d6b67bc0f6b35506f2a0d2a0bb3d59b77e19762a9bfd4234b88" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.924047 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-cjmn2" Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.948094 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" event={"ID":"14ff3b9a-634a-433c-a454-3d09f6fddbf0","Type":"ContainerStarted","Data":"9f5e0deb21fa04290066b2c4d6c914c6a47f0dfd82e4a59004cde764f25609a2"} Nov 25 10:06:25 crc kubenswrapper[4926]: I1125 10:06:25.977127 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" podStartSLOduration=2.977104257 podStartE2EDuration="2.977104257s" podCreationTimestamp="2025-11-25 10:06:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:25.971262881 +0000 UTC m=+1082.655219308" watchObservedRunningTime="2025-11-25 10:06:25.977104257 +0000 UTC m=+1082.661060684" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.007001 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d785fcd8-5cqwz"] Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.008598 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.033005 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.033177 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.033273 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wxp4z" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.033453 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.033571 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 10:06:26 crc kubenswrapper[4926]: I1125 10:06:26.038577 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d785fcd8-5cqwz"] Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.172354 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvpnx\" (UniqueName: \"kubernetes.io/projected/e84974e8-06ae-4d64-8a23-5054daa2a0e8-kube-api-access-mvpnx\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.172671 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-combined-ca-bundle\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.172793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e84974e8-06ae-4d64-8a23-5054daa2a0e8-logs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.172889 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-config-data\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.172976 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-internal-tls-certs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.173054 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-public-tls-certs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.173073 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-scripts\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.274845 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-public-tls-certs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.274885 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-scripts\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.274970 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvpnx\" (UniqueName: \"kubernetes.io/projected/e84974e8-06ae-4d64-8a23-5054daa2a0e8-kube-api-access-mvpnx\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.274987 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-combined-ca-bundle\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.275015 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e84974e8-06ae-4d64-8a23-5054daa2a0e8-logs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.275048 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-config-data\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.275079 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-internal-tls-certs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.276554 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e84974e8-06ae-4d64-8a23-5054daa2a0e8-logs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.280765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-internal-tls-certs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.280893 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-scripts\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.282356 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-combined-ca-bundle\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.283227 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-public-tls-certs\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.287483 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e84974e8-06ae-4d64-8a23-5054daa2a0e8-config-data\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.296383 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvpnx\" (UniqueName: \"kubernetes.io/projected/e84974e8-06ae-4d64-8a23-5054daa2a0e8-kube-api-access-mvpnx\") pod \"placement-d785fcd8-5cqwz\" (UID: \"e84974e8-06ae-4d64-8a23-5054daa2a0e8\") " pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.362582 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:26.960861 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:27.683621 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d785fcd8-5cqwz"] Nov 25 10:06:27 crc kubenswrapper[4926]: I1125 10:06:27.911744 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6cdc678f6f-htchg"] Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.819580 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.874808 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-combined-ca-bundle\") pod \"545481d3-3362-45f8-929b-71e8c75490dc\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.874959 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qx92\" (UniqueName: \"kubernetes.io/projected/545481d3-3362-45f8-929b-71e8c75490dc-kube-api-access-5qx92\") pod \"545481d3-3362-45f8-929b-71e8c75490dc\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.875021 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-scripts\") pod \"545481d3-3362-45f8-929b-71e8c75490dc\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.875101 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-config-data\") pod \"545481d3-3362-45f8-929b-71e8c75490dc\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.875157 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-fernet-keys\") pod \"545481d3-3362-45f8-929b-71e8c75490dc\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.875582 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-credential-keys\") pod \"545481d3-3362-45f8-929b-71e8c75490dc\" (UID: \"545481d3-3362-45f8-929b-71e8c75490dc\") " Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.880119 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "545481d3-3362-45f8-929b-71e8c75490dc" (UID: "545481d3-3362-45f8-929b-71e8c75490dc"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.880554 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/545481d3-3362-45f8-929b-71e8c75490dc-kube-api-access-5qx92" (OuterVolumeSpecName: "kube-api-access-5qx92") pod "545481d3-3362-45f8-929b-71e8c75490dc" (UID: "545481d3-3362-45f8-929b-71e8c75490dc"). InnerVolumeSpecName "kube-api-access-5qx92". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.881186 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "545481d3-3362-45f8-929b-71e8c75490dc" (UID: "545481d3-3362-45f8-929b-71e8c75490dc"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.881487 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-scripts" (OuterVolumeSpecName: "scripts") pod "545481d3-3362-45f8-929b-71e8c75490dc" (UID: "545481d3-3362-45f8-929b-71e8c75490dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.918337 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-config-data" (OuterVolumeSpecName: "config-data") pod "545481d3-3362-45f8-929b-71e8c75490dc" (UID: "545481d3-3362-45f8-929b-71e8c75490dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.918850 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "545481d3-3362-45f8-929b-71e8c75490dc" (UID: "545481d3-3362-45f8-929b-71e8c75490dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.977985 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.978026 4926 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.978040 4926 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.978053 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.978066 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qx92\" (UniqueName: \"kubernetes.io/projected/545481d3-3362-45f8-929b-71e8c75490dc-kube-api-access-5qx92\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:30 crc kubenswrapper[4926]: I1125 10:06:30.978078 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/545481d3-3362-45f8-929b-71e8c75490dc-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.009980 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-9fjls" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.010003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-9fjls" event={"ID":"545481d3-3362-45f8-929b-71e8c75490dc","Type":"ContainerDied","Data":"c01dfdf3f2e0a28206c87655a20c64a5631263c8d046784ae77ad10bfc7251ab"} Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.010051 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c01dfdf3f2e0a28206c87655a20c64a5631263c8d046784ae77ad10bfc7251ab" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.010877 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdc678f6f-htchg" event={"ID":"331a3347-d9a9-4a86-a552-58ab450406d4","Type":"ContainerStarted","Data":"9a90139e9fddd97f3fb355e44c414fc1070b6bc25ea40044ecbd6133217a0498"} Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.011667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d785fcd8-5cqwz" event={"ID":"e84974e8-06ae-4d64-8a23-5054daa2a0e8","Type":"ContainerStarted","Data":"bbbefd0e4ffb203c0be29a529c287a71459abf28a053e96e87cd1fb2f24cc174"} Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.963844 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-5c5477c6d9-t4dvb"] Nov 25 10:06:31 crc kubenswrapper[4926]: E1125 10:06:31.965006 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="545481d3-3362-45f8-929b-71e8c75490dc" containerName="keystone-bootstrap" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.965105 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="545481d3-3362-45f8-929b-71e8c75490dc" containerName="keystone-bootstrap" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.965423 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="545481d3-3362-45f8-929b-71e8c75490dc" containerName="keystone-bootstrap" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.966282 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.970424 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.970702 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.971015 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.971353 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.971602 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.971738 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c77m5" Nov 25 10:06:31 crc kubenswrapper[4926]: I1125 10:06:31.991547 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c5477c6d9-t4dvb"] Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.022843 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdc678f6f-htchg" event={"ID":"331a3347-d9a9-4a86-a552-58ab450406d4","Type":"ContainerStarted","Data":"e7f71c41e6ccd653120f23b4459a6edcb34d9eed7268b9b3ebbe7358e98ece25"} Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034589 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-config-data\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034654 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-public-tls-certs\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034681 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-fernet-keys\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034734 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-internal-tls-certs\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034779 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvxd6\" (UniqueName: \"kubernetes.io/projected/92634e60-5ef6-4047-ad6d-0d5a58f77d05-kube-api-access-gvxd6\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034805 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-scripts\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034849 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-combined-ca-bundle\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.034880 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-credential-keys\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136042 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-config-data\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136097 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-public-tls-certs\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136122 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-fernet-keys\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136157 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-internal-tls-certs\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136183 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvxd6\" (UniqueName: \"kubernetes.io/projected/92634e60-5ef6-4047-ad6d-0d5a58f77d05-kube-api-access-gvxd6\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136203 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-scripts\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136228 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-combined-ca-bundle\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.136247 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-credential-keys\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.139271 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-credential-keys\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.140762 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-public-tls-certs\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.141201 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-config-data\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.141789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-internal-tls-certs\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.143114 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-combined-ca-bundle\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.143629 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-scripts\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.161282 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/92634e60-5ef6-4047-ad6d-0d5a58f77d05-fernet-keys\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.170499 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvxd6\" (UniqueName: \"kubernetes.io/projected/92634e60-5ef6-4047-ad6d-0d5a58f77d05-kube-api-access-gvxd6\") pod \"keystone-5c5477c6d9-t4dvb\" (UID: \"92634e60-5ef6-4047-ad6d-0d5a58f77d05\") " pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.307517 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:06:32 crc kubenswrapper[4926]: I1125 10:06:32.733453 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-5c5477c6d9-t4dvb"] Nov 25 10:06:32 crc kubenswrapper[4926]: W1125 10:06:32.739637 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92634e60_5ef6_4047_ad6d_0d5a58f77d05.slice/crio-b30ee165c087c70613ebb63e18f012ea47d218036f090f83638de7e4764eab34 WatchSource:0}: Error finding container b30ee165c087c70613ebb63e18f012ea47d218036f090f83638de7e4764eab34: Status 404 returned error can't find the container with id b30ee165c087c70613ebb63e18f012ea47d218036f090f83638de7e4764eab34 Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.034657 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c5477c6d9-t4dvb" event={"ID":"92634e60-5ef6-4047-ad6d-0d5a58f77d05","Type":"ContainerStarted","Data":"654de40fdd73157f1a89295cc5975eaa8de9b7742ffe2a2c7739c12ab6bbaa74"} Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.035031 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-5c5477c6d9-t4dvb" event={"ID":"92634e60-5ef6-4047-ad6d-0d5a58f77d05","Type":"ContainerStarted","Data":"b30ee165c087c70613ebb63e18f012ea47d218036f090f83638de7e4764eab34"} Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.037690 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d785fcd8-5cqwz" event={"ID":"e84974e8-06ae-4d64-8a23-5054daa2a0e8","Type":"ContainerStarted","Data":"532b56e95f2bbd3aa340fa489558e5e59b20e88e83075058870d1ca4661f386c"} Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.037732 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d785fcd8-5cqwz" event={"ID":"e84974e8-06ae-4d64-8a23-5054daa2a0e8","Type":"ContainerStarted","Data":"a6896aa3fd5f04c63d58edd61e0e8c606eca698fc6e7095aae1b0aa4f4a14824"} Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.039113 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.039152 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.042673 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerStarted","Data":"8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d"} Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.044805 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6cdc678f6f-htchg" event={"ID":"331a3347-d9a9-4a86-a552-58ab450406d4","Type":"ContainerStarted","Data":"065b2d07891a7bc593d33512dcbd47e655dd8e0571c364a46a2c0e3da0ce4dcf"} Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.045860 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.072344 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-d785fcd8-5cqwz" podStartSLOduration=8.072323238 podStartE2EDuration="8.072323238s" podCreationTimestamp="2025-11-25 10:06:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:33.062528775 +0000 UTC m=+1089.746485242" watchObservedRunningTime="2025-11-25 10:06:33.072323238 +0000 UTC m=+1089.756279665" Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.085232 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6cdc678f6f-htchg" podStartSLOduration=8.085215688 podStartE2EDuration="8.085215688s" podCreationTimestamp="2025-11-25 10:06:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:33.084926991 +0000 UTC m=+1089.768883428" watchObservedRunningTime="2025-11-25 10:06:33.085215688 +0000 UTC m=+1089.769172115" Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.373146 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.443359 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5"] Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.448501 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerName="dnsmasq-dns" containerID="cri-o://1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c" gracePeriod=10 Nov 25 10:06:33 crc kubenswrapper[4926]: I1125 10:06:33.897520 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.053252 4926 generic.go:334] "Generic (PLEG): container finished" podID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerID="1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c" exitCode=0 Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.053338 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.053385 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" event={"ID":"b4001327-fff9-4fe5-9ae8-3a47dea56a57","Type":"ContainerDied","Data":"1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c"} Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.053412 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5" event={"ID":"b4001327-fff9-4fe5-9ae8-3a47dea56a57","Type":"ContainerDied","Data":"ccf39292d0fde2c6797e1daf7a324d06b320aec91e9ecbcb507a99fa7235d54d"} Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.053429 4926 scope.go:117] "RemoveContainer" containerID="1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.076865 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-dns-svc\") pod \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.076931 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-5c5477c6d9-t4dvb" podStartSLOduration=3.076904756 podStartE2EDuration="3.076904756s" podCreationTimestamp="2025-11-25 10:06:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:34.067968564 +0000 UTC m=+1090.751924991" watchObservedRunningTime="2025-11-25 10:06:34.076904756 +0000 UTC m=+1090.760861183" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.077722 4926 scope.go:117] "RemoveContainer" containerID="f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.076980 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-sb\") pod \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.077998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-nb\") pod \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.078060 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-config\") pod \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.078104 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9k9p7\" (UniqueName: \"kubernetes.io/projected/b4001327-fff9-4fe5-9ae8-3a47dea56a57-kube-api-access-9k9p7\") pod \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\" (UID: \"b4001327-fff9-4fe5-9ae8-3a47dea56a57\") " Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.086800 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4001327-fff9-4fe5-9ae8-3a47dea56a57-kube-api-access-9k9p7" (OuterVolumeSpecName: "kube-api-access-9k9p7") pod "b4001327-fff9-4fe5-9ae8-3a47dea56a57" (UID: "b4001327-fff9-4fe5-9ae8-3a47dea56a57"). InnerVolumeSpecName "kube-api-access-9k9p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.120373 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b4001327-fff9-4fe5-9ae8-3a47dea56a57" (UID: "b4001327-fff9-4fe5-9ae8-3a47dea56a57"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.124644 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b4001327-fff9-4fe5-9ae8-3a47dea56a57" (UID: "b4001327-fff9-4fe5-9ae8-3a47dea56a57"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.130492 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-config" (OuterVolumeSpecName: "config") pod "b4001327-fff9-4fe5-9ae8-3a47dea56a57" (UID: "b4001327-fff9-4fe5-9ae8-3a47dea56a57"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.139954 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b4001327-fff9-4fe5-9ae8-3a47dea56a57" (UID: "b4001327-fff9-4fe5-9ae8-3a47dea56a57"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.183381 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.184999 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.185030 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.185039 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4001327-fff9-4fe5-9ae8-3a47dea56a57-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.185049 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9k9p7\" (UniqueName: \"kubernetes.io/projected/b4001327-fff9-4fe5-9ae8-3a47dea56a57-kube-api-access-9k9p7\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.194489 4926 scope.go:117] "RemoveContainer" containerID="1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c" Nov 25 10:06:34 crc kubenswrapper[4926]: E1125 10:06:34.194967 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c\": container with ID starting with 1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c not found: ID does not exist" containerID="1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.194996 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c"} err="failed to get container status \"1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c\": rpc error: code = NotFound desc = could not find container \"1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c\": container with ID starting with 1104d55cbc3c008c38681c2ada7ab4c513dfddb13ac54e21f159371062610e5c not found: ID does not exist" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.195037 4926 scope.go:117] "RemoveContainer" containerID="f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559" Nov 25 10:06:34 crc kubenswrapper[4926]: E1125 10:06:34.195324 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559\": container with ID starting with f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559 not found: ID does not exist" containerID="f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.195375 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559"} err="failed to get container status \"f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559\": rpc error: code = NotFound desc = could not find container \"f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559\": container with ID starting with f825745122068f7089f944fd6cd3db57436fcd96b0e59b452a3449c4f7536559 not found: ID does not exist" Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.384742 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5"] Nov 25 10:06:34 crc kubenswrapper[4926]: I1125 10:06:34.391851 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b6dbdb6f5-k9ln5"] Nov 25 10:06:36 crc kubenswrapper[4926]: I1125 10:06:36.025989 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" path="/var/lib/kubelet/pods/b4001327-fff9-4fe5-9ae8-3a47dea56a57/volumes" Nov 25 10:06:37 crc kubenswrapper[4926]: I1125 10:06:37.099507 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k58gt" event={"ID":"af3d71a6-8937-4d95-9152-d77c1afc39eb","Type":"ContainerStarted","Data":"392003ba1866bcf1d975a622e4aae4a9d6bbe1ccb80a0561fbee4f4e281f25b4"} Nov 25 10:06:37 crc kubenswrapper[4926]: I1125 10:06:37.123262 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-k58gt" podStartSLOduration=2.474525062 podStartE2EDuration="40.123243727s" podCreationTimestamp="2025-11-25 10:05:57 +0000 UTC" firstStartedPulling="2025-11-25 10:05:58.241676098 +0000 UTC m=+1054.925632525" lastFinishedPulling="2025-11-25 10:06:35.890394763 +0000 UTC m=+1092.574351190" observedRunningTime="2025-11-25 10:06:37.115184227 +0000 UTC m=+1093.799140654" watchObservedRunningTime="2025-11-25 10:06:37.123243727 +0000 UTC m=+1093.807200154" Nov 25 10:06:39 crc kubenswrapper[4926]: I1125 10:06:39.116925 4926 generic.go:334] "Generic (PLEG): container finished" podID="af3d71a6-8937-4d95-9152-d77c1afc39eb" containerID="392003ba1866bcf1d975a622e4aae4a9d6bbe1ccb80a0561fbee4f4e281f25b4" exitCode=0 Nov 25 10:06:39 crc kubenswrapper[4926]: I1125 10:06:39.117013 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k58gt" event={"ID":"af3d71a6-8937-4d95-9152-d77c1afc39eb","Type":"ContainerDied","Data":"392003ba1866bcf1d975a622e4aae4a9d6bbe1ccb80a0561fbee4f4e281f25b4"} Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.665445 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k58gt" Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.793892 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-combined-ca-bundle\") pod \"af3d71a6-8937-4d95-9152-d77c1afc39eb\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.793985 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9crcl\" (UniqueName: \"kubernetes.io/projected/af3d71a6-8937-4d95-9152-d77c1afc39eb-kube-api-access-9crcl\") pod \"af3d71a6-8937-4d95-9152-d77c1afc39eb\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.794084 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-db-sync-config-data\") pod \"af3d71a6-8937-4d95-9152-d77c1afc39eb\" (UID: \"af3d71a6-8937-4d95-9152-d77c1afc39eb\") " Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.798671 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af3d71a6-8937-4d95-9152-d77c1afc39eb-kube-api-access-9crcl" (OuterVolumeSpecName: "kube-api-access-9crcl") pod "af3d71a6-8937-4d95-9152-d77c1afc39eb" (UID: "af3d71a6-8937-4d95-9152-d77c1afc39eb"). InnerVolumeSpecName "kube-api-access-9crcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.799196 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "af3d71a6-8937-4d95-9152-d77c1afc39eb" (UID: "af3d71a6-8937-4d95-9152-d77c1afc39eb"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.821024 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af3d71a6-8937-4d95-9152-d77c1afc39eb" (UID: "af3d71a6-8937-4d95-9152-d77c1afc39eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.896465 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.896503 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9crcl\" (UniqueName: \"kubernetes.io/projected/af3d71a6-8937-4d95-9152-d77c1afc39eb-kube-api-access-9crcl\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:40 crc kubenswrapper[4926]: I1125 10:06:40.896517 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/af3d71a6-8937-4d95-9152-d77c1afc39eb-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.141443 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-k58gt" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.141531 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-k58gt" event={"ID":"af3d71a6-8937-4d95-9152-d77c1afc39eb","Type":"ContainerDied","Data":"5fc08b2d5d6de8c4d0abe67f4df52ed55bf3adbd442321c150b3ae68b50cf125"} Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.141565 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fc08b2d5d6de8c4d0abe67f4df52ed55bf3adbd442321c150b3ae68b50cf125" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.144595 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerStarted","Data":"0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df"} Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.144909 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="proxy-httpd" containerID="cri-o://0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df" gracePeriod=30 Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.144907 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-central-agent" containerID="cri-o://5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5" gracePeriod=30 Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.145172 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-notification-agent" containerID="cri-o://e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442" gracePeriod=30 Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.145200 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="sg-core" containerID="cri-o://8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d" gracePeriod=30 Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.180197 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.954957414 podStartE2EDuration="44.180177186s" podCreationTimestamp="2025-11-25 10:05:57 +0000 UTC" firstStartedPulling="2025-11-25 10:05:58.44600368 +0000 UTC m=+1055.129960107" lastFinishedPulling="2025-11-25 10:06:40.671223452 +0000 UTC m=+1097.355179879" observedRunningTime="2025-11-25 10:06:41.174491385 +0000 UTC m=+1097.858447812" watchObservedRunningTime="2025-11-25 10:06:41.180177186 +0000 UTC m=+1097.864133613" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.408268 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-75b9f64b55-mrc6w"] Nov 25 10:06:41 crc kubenswrapper[4926]: E1125 10:06:41.408674 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af3d71a6-8937-4d95-9152-d77c1afc39eb" containerName="barbican-db-sync" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.408694 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af3d71a6-8937-4d95-9152-d77c1afc39eb" containerName="barbican-db-sync" Nov 25 10:06:41 crc kubenswrapper[4926]: E1125 10:06:41.408712 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerName="init" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.408719 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerName="init" Nov 25 10:06:41 crc kubenswrapper[4926]: E1125 10:06:41.408734 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerName="dnsmasq-dns" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.408742 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerName="dnsmasq-dns" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.408990 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af3d71a6-8937-4d95-9152-d77c1afc39eb" containerName="barbican-db-sync" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.409016 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4001327-fff9-4fe5-9ae8-3a47dea56a57" containerName="dnsmasq-dns" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.410033 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.413835 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.419873 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.420884 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fc6gp" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.423054 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5dd564f876-448mb"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.424637 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.427457 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.446165 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5dd564f876-448mb"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.454804 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-75b9f64b55-mrc6w"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.527892 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-869f779d85-wffhz"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.529890 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.571412 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-wffhz"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609040 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-combined-ca-bundle\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609092 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-combined-ca-bundle\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609129 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-config-data\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609159 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glvsv\" (UniqueName: \"kubernetes.io/projected/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-kube-api-access-glvsv\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609202 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-config-data\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609259 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhmvn\" (UniqueName: \"kubernetes.io/projected/a33145e6-61bd-4706-b9cf-bf2e59906a6b-kube-api-access-nhmvn\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609292 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-logs\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609332 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a33145e6-61bd-4706-b9cf-bf2e59906a6b-logs\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609356 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-config-data-custom\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.609392 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-config-data-custom\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.708561 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-58855fbcfd-9r5c5"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712425 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhmvn\" (UniqueName: \"kubernetes.io/projected/a33145e6-61bd-4706-b9cf-bf2e59906a6b-kube-api-access-nhmvn\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712474 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-dns-svc\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712502 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712563 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-logs\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712598 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a33145e6-61bd-4706-b9cf-bf2e59906a6b-logs\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712620 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-config-data-custom\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712662 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-config-data-custom\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712693 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gskcw\" (UniqueName: \"kubernetes.io/projected/0ffe198c-1848-458f-988f-d8df3e03f628-kube-api-access-gskcw\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712711 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-combined-ca-bundle\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712735 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-combined-ca-bundle\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712752 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712781 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-config-data\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glvsv\" (UniqueName: \"kubernetes.io/projected/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-kube-api-access-glvsv\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712827 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-config\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.712853 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-config-data\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.714006 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-logs\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.714296 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a33145e6-61bd-4706-b9cf-bf2e59906a6b-logs\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.721488 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-combined-ca-bundle\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.728828 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-config-data\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.729722 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.733194 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-config-data-custom\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.733324 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.734731 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a33145e6-61bd-4706-b9cf-bf2e59906a6b-config-data-custom\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.735584 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-combined-ca-bundle\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.740281 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-config-data\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.744381 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58855fbcfd-9r5c5"] Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.750577 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glvsv\" (UniqueName: \"kubernetes.io/projected/b069a045-2f4e-44f9-8d85-cdacf9bb1ccd-kube-api-access-glvsv\") pod \"barbican-worker-75b9f64b55-mrc6w\" (UID: \"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd\") " pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.751363 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhmvn\" (UniqueName: \"kubernetes.io/projected/a33145e6-61bd-4706-b9cf-bf2e59906a6b-kube-api-access-nhmvn\") pod \"barbican-keystone-listener-5dd564f876-448mb\" (UID: \"a33145e6-61bd-4706-b9cf-bf2e59906a6b\") " pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.764729 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5dd564f876-448mb" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.817465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.817534 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-config\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.817616 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-dns-svc\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.817634 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.817709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gskcw\" (UniqueName: \"kubernetes.io/projected/0ffe198c-1848-458f-988f-d8df3e03f628-kube-api-access-gskcw\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.818732 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-sb\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.819334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-config\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.820033 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-nb\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.821013 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-dns-svc\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.836756 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gskcw\" (UniqueName: \"kubernetes.io/projected/0ffe198c-1848-458f-988f-d8df3e03f628-kube-api-access-gskcw\") pod \"dnsmasq-dns-869f779d85-wffhz\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.871504 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.919712 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9083ac42-380c-471a-a89f-27c6318e5493-logs\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.920058 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5jng\" (UniqueName: \"kubernetes.io/projected/9083ac42-380c-471a-a89f-27c6318e5493-kube-api-access-l5jng\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.920305 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-combined-ca-bundle\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.920557 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data-custom\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:41 crc kubenswrapper[4926]: I1125 10:06:41.920611 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.022449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data-custom\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.022504 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.022528 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9083ac42-380c-471a-a89f-27c6318e5493-logs\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.022561 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5jng\" (UniqueName: \"kubernetes.io/projected/9083ac42-380c-471a-a89f-27c6318e5493-kube-api-access-l5jng\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.022650 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-combined-ca-bundle\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.027468 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9083ac42-380c-471a-a89f-27c6318e5493-logs\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.029708 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-75b9f64b55-mrc6w" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.043163 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-combined-ca-bundle\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.043565 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.045009 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data-custom\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.048031 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5jng\" (UniqueName: \"kubernetes.io/projected/9083ac42-380c-471a-a89f-27c6318e5493-kube-api-access-l5jng\") pod \"barbican-api-58855fbcfd-9r5c5\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.156919 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k6rtr" event={"ID":"65e6c095-3a19-4536-b55a-c79c991badd0","Type":"ContainerStarted","Data":"96b38b7ef994115e7ebb8e7cf6948d5b45a0faee4a0cdb3c89f526104b81337d"} Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.161568 4926 generic.go:334] "Generic (PLEG): container finished" podID="170174f9-d802-42f5-97a1-13c6bbe50460" containerID="0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df" exitCode=0 Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.161599 4926 generic.go:334] "Generic (PLEG): container finished" podID="170174f9-d802-42f5-97a1-13c6bbe50460" containerID="8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d" exitCode=2 Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.161608 4926 generic.go:334] "Generic (PLEG): container finished" podID="170174f9-d802-42f5-97a1-13c6bbe50460" containerID="5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5" exitCode=0 Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.161627 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerDied","Data":"0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df"} Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.161650 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerDied","Data":"8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d"} Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.161661 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerDied","Data":"5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5"} Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.198238 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-k6rtr" podStartSLOduration=3.394347942 podStartE2EDuration="46.198211698s" podCreationTimestamp="2025-11-25 10:05:56 +0000 UTC" firstStartedPulling="2025-11-25 10:05:57.864530945 +0000 UTC m=+1054.548487372" lastFinishedPulling="2025-11-25 10:06:40.668394661 +0000 UTC m=+1097.352351128" observedRunningTime="2025-11-25 10:06:42.18016409 +0000 UTC m=+1098.864120527" watchObservedRunningTime="2025-11-25 10:06:42.198211698 +0000 UTC m=+1098.882168125" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.255408 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.306150 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5dd564f876-448mb"] Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.386920 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-wffhz"] Nov 25 10:06:42 crc kubenswrapper[4926]: W1125 10:06:42.395502 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ffe198c_1848_458f_988f_d8df3e03f628.slice/crio-0ad687fd03407350f3d3be9d012b503aa07d112b06529da4581db4e2c3cdbf3d WatchSource:0}: Error finding container 0ad687fd03407350f3d3be9d012b503aa07d112b06529da4581db4e2c3cdbf3d: Status 404 returned error can't find the container with id 0ad687fd03407350f3d3be9d012b503aa07d112b06529da4581db4e2c3cdbf3d Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.514006 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-75b9f64b55-mrc6w"] Nov 25 10:06:42 crc kubenswrapper[4926]: W1125 10:06:42.516463 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb069a045_2f4e_44f9_8d85_cdacf9bb1ccd.slice/crio-5466c440b003b7c6057d88b7cf34e5df44fd8e4021db37ef6ce625481fa98728 WatchSource:0}: Error finding container 5466c440b003b7c6057d88b7cf34e5df44fd8e4021db37ef6ce625481fa98728: Status 404 returned error can't find the container with id 5466c440b003b7c6057d88b7cf34e5df44fd8e4021db37ef6ce625481fa98728 Nov 25 10:06:42 crc kubenswrapper[4926]: I1125 10:06:42.705255 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-58855fbcfd-9r5c5"] Nov 25 10:06:42 crc kubenswrapper[4926]: W1125 10:06:42.727836 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9083ac42_380c_471a_a89f_27c6318e5493.slice/crio-6ce0568bf33422a9fff2c9de5a457ae044666be6c20c889a6b7618ef2f2ff3ae WatchSource:0}: Error finding container 6ce0568bf33422a9fff2c9de5a457ae044666be6c20c889a6b7618ef2f2ff3ae: Status 404 returned error can't find the container with id 6ce0568bf33422a9fff2c9de5a457ae044666be6c20c889a6b7618ef2f2ff3ae Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.171888 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75b9f64b55-mrc6w" event={"ID":"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd","Type":"ContainerStarted","Data":"5466c440b003b7c6057d88b7cf34e5df44fd8e4021db37ef6ce625481fa98728"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.173377 4926 generic.go:334] "Generic (PLEG): container finished" podID="0ffe198c-1848-458f-988f-d8df3e03f628" containerID="3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53" exitCode=0 Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.173440 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-wffhz" event={"ID":"0ffe198c-1848-458f-988f-d8df3e03f628","Type":"ContainerDied","Data":"3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.173493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-wffhz" event={"ID":"0ffe198c-1848-458f-988f-d8df3e03f628","Type":"ContainerStarted","Data":"0ad687fd03407350f3d3be9d012b503aa07d112b06529da4581db4e2c3cdbf3d"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.179226 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dd564f876-448mb" event={"ID":"a33145e6-61bd-4706-b9cf-bf2e59906a6b","Type":"ContainerStarted","Data":"1760ffd9af9574dcebbe05aaabe9190afe9dbffa0096a9b1fb712a4357d6f67c"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.183781 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58855fbcfd-9r5c5" event={"ID":"9083ac42-380c-471a-a89f-27c6318e5493","Type":"ContainerStarted","Data":"a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.183829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58855fbcfd-9r5c5" event={"ID":"9083ac42-380c-471a-a89f-27c6318e5493","Type":"ContainerStarted","Data":"4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.183839 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58855fbcfd-9r5c5" event={"ID":"9083ac42-380c-471a-a89f-27c6318e5493","Type":"ContainerStarted","Data":"6ce0568bf33422a9fff2c9de5a457ae044666be6c20c889a6b7618ef2f2ff3ae"} Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.184045 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:43 crc kubenswrapper[4926]: I1125 10:06:43.253327 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-58855fbcfd-9r5c5" podStartSLOduration=2.253311369 podStartE2EDuration="2.253311369s" podCreationTimestamp="2025-11-25 10:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:43.251199636 +0000 UTC m=+1099.935156063" watchObservedRunningTime="2025-11-25 10:06:43.253311369 +0000 UTC m=+1099.937267796" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.021366 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.021668 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.191754 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.440306 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-8594cdbfd6-wcgzz"] Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.442661 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.445341 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.445613 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.473609 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8594cdbfd6-wcgzz"] Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478491 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-internal-tls-certs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478579 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-public-tls-certs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478645 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-combined-ca-bundle\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478672 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-config-data-custom\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478723 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vw8nk\" (UniqueName: \"kubernetes.io/projected/a2920249-1629-4833-a8a0-0b5fee1c2303-kube-api-access-vw8nk\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2920249-1629-4833-a8a0-0b5fee1c2303-logs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.478973 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-config-data\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.579477 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2920249-1629-4833-a8a0-0b5fee1c2303-logs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.579817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-config-data\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.579852 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-internal-tls-certs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.579869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2920249-1629-4833-a8a0-0b5fee1c2303-logs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.579883 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-public-tls-certs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.579972 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-combined-ca-bundle\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.580017 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-config-data-custom\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.580087 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vw8nk\" (UniqueName: \"kubernetes.io/projected/a2920249-1629-4833-a8a0-0b5fee1c2303-kube-api-access-vw8nk\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.583437 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-combined-ca-bundle\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.583692 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-internal-tls-certs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.584483 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-config-data-custom\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.584803 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-public-tls-certs\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.586008 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2920249-1629-4833-a8a0-0b5fee1c2303-config-data\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.597220 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vw8nk\" (UniqueName: \"kubernetes.io/projected/a2920249-1629-4833-a8a0-0b5fee1c2303-kube-api-access-vw8nk\") pod \"barbican-api-8594cdbfd6-wcgzz\" (UID: \"a2920249-1629-4833-a8a0-0b5fee1c2303\") " pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:44 crc kubenswrapper[4926]: I1125 10:06:44.779482 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.200802 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dd564f876-448mb" event={"ID":"a33145e6-61bd-4706-b9cf-bf2e59906a6b","Type":"ContainerStarted","Data":"9e9d7431c89af65db25c298ba4aaa34722064cd7055730245bffca414a838f69"} Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.201196 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5dd564f876-448mb" event={"ID":"a33145e6-61bd-4706-b9cf-bf2e59906a6b","Type":"ContainerStarted","Data":"0536547309f437f39519dea3e9eb82c9507a9c6fe7dcd31ceadfaefe6d30ea36"} Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.203151 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75b9f64b55-mrc6w" event={"ID":"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd","Type":"ContainerStarted","Data":"9837e29b495e8da829d5a2115b46d07adca81abdc0f94ed4b8d3a9ce2667ec11"} Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.203200 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-75b9f64b55-mrc6w" event={"ID":"b069a045-2f4e-44f9-8d85-cdacf9bb1ccd","Type":"ContainerStarted","Data":"8366cd4d2f7c685dc6b1ecb6dd5ac7111c8d070c3280592cc1ebcfac819440ea"} Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.205786 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-wffhz" event={"ID":"0ffe198c-1848-458f-988f-d8df3e03f628","Type":"ContainerStarted","Data":"8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240"} Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.241694 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5dd564f876-448mb" podStartSLOduration=2.367080254 podStartE2EDuration="4.241659418s" podCreationTimestamp="2025-11-25 10:06:41 +0000 UTC" firstStartedPulling="2025-11-25 10:06:42.32238567 +0000 UTC m=+1099.006342097" lastFinishedPulling="2025-11-25 10:06:44.196964834 +0000 UTC m=+1100.880921261" observedRunningTime="2025-11-25 10:06:45.218772749 +0000 UTC m=+1101.902729206" watchObservedRunningTime="2025-11-25 10:06:45.241659418 +0000 UTC m=+1101.925615865" Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.245342 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-75b9f64b55-mrc6w" podStartSLOduration=2.56708816 podStartE2EDuration="4.245327199s" podCreationTimestamp="2025-11-25 10:06:41 +0000 UTC" firstStartedPulling="2025-11-25 10:06:42.518742375 +0000 UTC m=+1099.202698802" lastFinishedPulling="2025-11-25 10:06:44.196981414 +0000 UTC m=+1100.880937841" observedRunningTime="2025-11-25 10:06:45.235251459 +0000 UTC m=+1101.919207886" watchObservedRunningTime="2025-11-25 10:06:45.245327199 +0000 UTC m=+1101.929283626" Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.257851 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-869f779d85-wffhz" podStartSLOduration=4.257833219 podStartE2EDuration="4.257833219s" podCreationTimestamp="2025-11-25 10:06:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:45.252383054 +0000 UTC m=+1101.936339481" watchObservedRunningTime="2025-11-25 10:06:45.257833219 +0000 UTC m=+1101.941789646" Nov 25 10:06:45 crc kubenswrapper[4926]: I1125 10:06:45.287183 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-8594cdbfd6-wcgzz"] Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.238801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8594cdbfd6-wcgzz" event={"ID":"a2920249-1629-4833-a8a0-0b5fee1c2303","Type":"ContainerStarted","Data":"238b7c6b658083f3c18eb05118236080bcc0828e12eb123a7fccb594d46a6213"} Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.239716 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8594cdbfd6-wcgzz" event={"ID":"a2920249-1629-4833-a8a0-0b5fee1c2303","Type":"ContainerStarted","Data":"87af974539a171418d6dc30adf067aa5be0d71d406ebbced58af27ff2e48aac1"} Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.239738 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-8594cdbfd6-wcgzz" event={"ID":"a2920249-1629-4833-a8a0-0b5fee1c2303","Type":"ContainerStarted","Data":"ecea1f142b18bdf4eff54f43149966e3b36b60acf955eeba4f1d650a6096f4af"} Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.240959 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.240999 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.241020 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.298852 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-8594cdbfd6-wcgzz" podStartSLOduration=2.298826671 podStartE2EDuration="2.298826671s" podCreationTimestamp="2025-11-25 10:06:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:46.294419941 +0000 UTC m=+1102.978376378" watchObservedRunningTime="2025-11-25 10:06:46.298826671 +0000 UTC m=+1102.982783108" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.641549 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.820012 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-run-httpd\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.820558 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.820704 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-sg-core-conf-yaml\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.821370 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-config-data\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.821413 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-combined-ca-bundle\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.821473 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfvrg\" (UniqueName: \"kubernetes.io/projected/170174f9-d802-42f5-97a1-13c6bbe50460-kube-api-access-zfvrg\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.821531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-log-httpd\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.821572 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-scripts\") pod \"170174f9-d802-42f5-97a1-13c6bbe50460\" (UID: \"170174f9-d802-42f5-97a1-13c6bbe50460\") " Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.822053 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.822066 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.826367 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-scripts" (OuterVolumeSpecName: "scripts") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.827636 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/170174f9-d802-42f5-97a1-13c6bbe50460-kube-api-access-zfvrg" (OuterVolumeSpecName: "kube-api-access-zfvrg") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "kube-api-access-zfvrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.856346 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.891149 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.911540 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-config-data" (OuterVolumeSpecName: "config-data") pod "170174f9-d802-42f5-97a1-13c6bbe50460" (UID: "170174f9-d802-42f5-97a1-13c6bbe50460"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.923847 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.923879 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.923892 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.923901 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/170174f9-d802-42f5-97a1-13c6bbe50460-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.923909 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfvrg\" (UniqueName: \"kubernetes.io/projected/170174f9-d802-42f5-97a1-13c6bbe50460-kube-api-access-zfvrg\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:46 crc kubenswrapper[4926]: I1125 10:06:46.923928 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/170174f9-d802-42f5-97a1-13c6bbe50460-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.248430 4926 generic.go:334] "Generic (PLEG): container finished" podID="65e6c095-3a19-4536-b55a-c79c991badd0" containerID="96b38b7ef994115e7ebb8e7cf6948d5b45a0faee4a0cdb3c89f526104b81337d" exitCode=0 Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.248537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k6rtr" event={"ID":"65e6c095-3a19-4536-b55a-c79c991badd0","Type":"ContainerDied","Data":"96b38b7ef994115e7ebb8e7cf6948d5b45a0faee4a0cdb3c89f526104b81337d"} Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.251877 4926 generic.go:334] "Generic (PLEG): container finished" podID="170174f9-d802-42f5-97a1-13c6bbe50460" containerID="e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442" exitCode=0 Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.251980 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerDied","Data":"e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442"} Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.252011 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.252077 4926 scope.go:117] "RemoveContainer" containerID="0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.252059 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"170174f9-d802-42f5-97a1-13c6bbe50460","Type":"ContainerDied","Data":"07c526368b5f6cfe3724a2306cf45413dc7ab120df963117a94928fec723df46"} Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.296872 4926 scope.go:117] "RemoveContainer" containerID="8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.306849 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.324274 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.346630 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.347027 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="proxy-httpd" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347042 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="proxy-httpd" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.347061 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="sg-core" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347066 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="sg-core" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.347080 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-notification-agent" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347087 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-notification-agent" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.347108 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-central-agent" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347115 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-central-agent" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347280 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="proxy-httpd" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347296 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-notification-agent" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347303 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="ceilometer-central-agent" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.347321 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" containerName="sg-core" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.349122 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.353418 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.354932 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.360143 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.362189 4926 scope.go:117] "RemoveContainer" containerID="e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.421864 4926 scope.go:117] "RemoveContainer" containerID="5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.442838 4926 scope.go:117] "RemoveContainer" containerID="0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.443269 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df\": container with ID starting with 0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df not found: ID does not exist" containerID="0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.443307 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df"} err="failed to get container status \"0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df\": rpc error: code = NotFound desc = could not find container \"0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df\": container with ID starting with 0338f3dfa80433ec189fb16b290ad8f50936dcf2b499897fb3366bff9a6268df not found: ID does not exist" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.443331 4926 scope.go:117] "RemoveContainer" containerID="8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.443669 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d\": container with ID starting with 8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d not found: ID does not exist" containerID="8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.443699 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d"} err="failed to get container status \"8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d\": rpc error: code = NotFound desc = could not find container \"8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d\": container with ID starting with 8fe39d51f348a469d3038ab0bb2dce197b37514e27c94ee8f73450f5eb58883d not found: ID does not exist" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.443724 4926 scope.go:117] "RemoveContainer" containerID="e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.443951 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442\": container with ID starting with e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442 not found: ID does not exist" containerID="e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.443976 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442"} err="failed to get container status \"e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442\": rpc error: code = NotFound desc = could not find container \"e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442\": container with ID starting with e7d196da2ac0f08fd000ba19f24791e4007406df9d8f9ffee04fe1a18b346442 not found: ID does not exist" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.443989 4926 scope.go:117] "RemoveContainer" containerID="5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5" Nov 25 10:06:47 crc kubenswrapper[4926]: E1125 10:06:47.444187 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5\": container with ID starting with 5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5 not found: ID does not exist" containerID="5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.444208 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5"} err="failed to get container status \"5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5\": rpc error: code = NotFound desc = could not find container \"5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5\": container with ID starting with 5f738505895e7bccf4494c610b50e0acd67402b4ef66e393e082ba71afb904a5 not found: ID does not exist" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.535790 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-log-httpd\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.535877 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.536059 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.536121 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-config-data\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.536172 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-scripts\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.536216 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-run-httpd\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.536387 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdl7m\" (UniqueName: \"kubernetes.io/projected/d1f2720a-802e-404e-9a13-a81db6466073-kube-api-access-tdl7m\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638312 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-run-httpd\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638444 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdl7m\" (UniqueName: \"kubernetes.io/projected/d1f2720a-802e-404e-9a13-a81db6466073-kube-api-access-tdl7m\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638523 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-log-httpd\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638635 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638695 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-config-data\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.638766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-scripts\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.639541 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-run-httpd\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.639683 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-log-httpd\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.645880 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-config-data\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.646668 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.663549 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.665171 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-scripts\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.671456 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdl7m\" (UniqueName: \"kubernetes.io/projected/d1f2720a-802e-404e-9a13-a81db6466073-kube-api-access-tdl7m\") pod \"ceilometer-0\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " pod="openstack/ceilometer-0" Nov 25 10:06:47 crc kubenswrapper[4926]: I1125 10:06:47.700092 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.024785 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="170174f9-d802-42f5-97a1-13c6bbe50460" path="/var/lib/kubelet/pods/170174f9-d802-42f5-97a1-13c6bbe50460/volumes" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.126250 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:06:48 crc kubenswrapper[4926]: W1125 10:06:48.131090 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd1f2720a_802e_404e_9a13_a81db6466073.slice/crio-4266b51673f65de180aaa17891dcc0a449ab3e32c8a15da54758cb5033a739b3 WatchSource:0}: Error finding container 4266b51673f65de180aaa17891dcc0a449ab3e32c8a15da54758cb5033a739b3: Status 404 returned error can't find the container with id 4266b51673f65de180aaa17891dcc0a449ab3e32c8a15da54758cb5033a739b3 Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.262571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerStarted","Data":"4266b51673f65de180aaa17891dcc0a449ab3e32c8a15da54758cb5033a739b3"} Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.640680 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.655143 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-config-data\") pod \"65e6c095-3a19-4536-b55a-c79c991badd0\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.655229 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6c095-3a19-4536-b55a-c79c991badd0-etc-machine-id\") pod \"65e6c095-3a19-4536-b55a-c79c991badd0\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.655274 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-scripts\") pod \"65e6c095-3a19-4536-b55a-c79c991badd0\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.655347 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwtsj\" (UniqueName: \"kubernetes.io/projected/65e6c095-3a19-4536-b55a-c79c991badd0-kube-api-access-hwtsj\") pod \"65e6c095-3a19-4536-b55a-c79c991badd0\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.655374 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-db-sync-config-data\") pod \"65e6c095-3a19-4536-b55a-c79c991badd0\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.655403 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-combined-ca-bundle\") pod \"65e6c095-3a19-4536-b55a-c79c991badd0\" (UID: \"65e6c095-3a19-4536-b55a-c79c991badd0\") " Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.657752 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/65e6c095-3a19-4536-b55a-c79c991badd0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "65e6c095-3a19-4536-b55a-c79c991badd0" (UID: "65e6c095-3a19-4536-b55a-c79c991badd0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.663267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-scripts" (OuterVolumeSpecName: "scripts") pod "65e6c095-3a19-4536-b55a-c79c991badd0" (UID: "65e6c095-3a19-4536-b55a-c79c991badd0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.665853 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65e6c095-3a19-4536-b55a-c79c991badd0-kube-api-access-hwtsj" (OuterVolumeSpecName: "kube-api-access-hwtsj") pod "65e6c095-3a19-4536-b55a-c79c991badd0" (UID: "65e6c095-3a19-4536-b55a-c79c991badd0"). InnerVolumeSpecName "kube-api-access-hwtsj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.668066 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "65e6c095-3a19-4536-b55a-c79c991badd0" (UID: "65e6c095-3a19-4536-b55a-c79c991badd0"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.708889 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "65e6c095-3a19-4536-b55a-c79c991badd0" (UID: "65e6c095-3a19-4536-b55a-c79c991badd0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.723977 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-config-data" (OuterVolumeSpecName: "config-data") pod "65e6c095-3a19-4536-b55a-c79c991badd0" (UID: "65e6c095-3a19-4536-b55a-c79c991badd0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.757183 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.757216 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/65e6c095-3a19-4536-b55a-c79c991badd0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.757226 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.757234 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwtsj\" (UniqueName: \"kubernetes.io/projected/65e6c095-3a19-4536-b55a-c79c991badd0-kube-api-access-hwtsj\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.757243 4926 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.757252 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65e6c095-3a19-4536-b55a-c79c991badd0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:48 crc kubenswrapper[4926]: I1125 10:06:48.796835 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.271331 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-k6rtr" event={"ID":"65e6c095-3a19-4536-b55a-c79c991badd0","Type":"ContainerDied","Data":"2f21aa166b9c7c23a51d8f906b99dce0e46c696c0b14bb1fe88bc97640ae40ec"} Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.271620 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f21aa166b9c7c23a51d8f906b99dce0e46c696c0b14bb1fe88bc97640ae40ec" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.271675 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-k6rtr" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.276076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerStarted","Data":"68f633a1421878cb98b424fad8270c42f7e0f2fcf2e3d73230c72d505c3085c5"} Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.534858 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:06:49 crc kubenswrapper[4926]: E1125 10:06:49.535235 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65e6c095-3a19-4536-b55a-c79c991badd0" containerName="cinder-db-sync" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.535251 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="65e6c095-3a19-4536-b55a-c79c991badd0" containerName="cinder-db-sync" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.535393 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="65e6c095-3a19-4536-b55a-c79c991badd0" containerName="cinder-db-sync" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.536283 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.540430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.540524 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nzjdg" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.540646 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.548760 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.564473 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.655214 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-wffhz"] Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.655755 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-869f779d85-wffhz" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" containerName="dnsmasq-dns" containerID="cri-o://8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240" gracePeriod=10 Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.657753 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.681484 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-scripts\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.681528 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jkw4\" (UniqueName: \"kubernetes.io/projected/1ca6a1d2-b396-44e8-9701-f3dd74b67724-kube-api-access-4jkw4\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.681555 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.681624 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.685833 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ca6a1d2-b396-44e8-9701-f3dd74b67724-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.685895 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.716797 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-pz6nn"] Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.723685 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.732256 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-pz6nn"] Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.791542 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.791603 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ca6a1d2-b396-44e8-9701-f3dd74b67724-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.791646 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.791728 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-scripts\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.791746 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jkw4\" (UniqueName: \"kubernetes.io/projected/1ca6a1d2-b396-44e8-9701-f3dd74b67724-kube-api-access-4jkw4\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.791771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.795086 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ca6a1d2-b396-44e8-9701-f3dd74b67724-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.804665 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.807161 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.809096 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.809479 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-scripts\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.854752 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jkw4\" (UniqueName: \"kubernetes.io/projected/1ca6a1d2-b396-44e8-9701-f3dd74b67724-kube-api-access-4jkw4\") pod \"cinder-scheduler-0\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.879156 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.881324 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.882417 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.891716 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.892788 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.892843 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcw27\" (UniqueName: \"kubernetes.io/projected/44037126-caa3-43d2-abf4-937f10d8fe2a-kube-api-access-tcw27\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.892872 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-dns-svc\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.892966 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-config\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.892991 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.901341 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994519 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994592 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-58nbk\" (UniqueName: \"kubernetes.io/projected/0405b7fb-14c1-4958-a25a-560bfd8bf859-kube-api-access-58nbk\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994652 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcw27\" (UniqueName: \"kubernetes.io/projected/44037126-caa3-43d2-abf4-937f10d8fe2a-kube-api-access-tcw27\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994707 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-dns-svc\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994754 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994782 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-scripts\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994856 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0405b7fb-14c1-4958-a25a-560bfd8bf859-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994888 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-config\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994916 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.994967 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0405b7fb-14c1-4958-a25a-560bfd8bf859-logs\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.995015 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.995044 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data-custom\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:49 crc kubenswrapper[4926]: I1125 10:06:49.996141 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-nb\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:49.996924 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-dns-svc\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:49.997173 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-sb\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.015443 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-config\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.021654 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcw27\" (UniqueName: \"kubernetes.io/projected/44037126-caa3-43d2-abf4-937f10d8fe2a-kube-api-access-tcw27\") pod \"dnsmasq-dns-58db5546cc-pz6nn\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.057500 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096441 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096493 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data-custom\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096574 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-58nbk\" (UniqueName: \"kubernetes.io/projected/0405b7fb-14c1-4958-a25a-560bfd8bf859-kube-api-access-58nbk\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096666 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096684 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-scripts\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0405b7fb-14c1-4958-a25a-560bfd8bf859-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.096756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0405b7fb-14c1-4958-a25a-560bfd8bf859-logs\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.097517 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0405b7fb-14c1-4958-a25a-560bfd8bf859-logs\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.098168 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0405b7fb-14c1-4958-a25a-560bfd8bf859-etc-machine-id\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.104479 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.108738 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-scripts\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.112118 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.114869 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data-custom\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.122233 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-58nbk\" (UniqueName: \"kubernetes.io/projected/0405b7fb-14c1-4958-a25a-560bfd8bf859-kube-api-access-58nbk\") pod \"cinder-api-0\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.246988 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:06:50 crc kubenswrapper[4926]: W1125 10:06:50.265922 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ca6a1d2_b396_44e8_9701_f3dd74b67724.slice/crio-c1b7158217a4729312096d76c3b83c4b98bcc010a536ec92a57435302afd4fd1 WatchSource:0}: Error finding container c1b7158217a4729312096d76c3b83c4b98bcc010a536ec92a57435302afd4fd1: Status 404 returned error can't find the container with id c1b7158217a4729312096d76c3b83c4b98bcc010a536ec92a57435302afd4fd1 Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.279626 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.309376 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerStarted","Data":"b7c5749cefc782238ff9e36132c6c640b89c0635cfa61e518ca3d56aeddb4bfb"} Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.311218 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ca6a1d2-b396-44e8-9701-f3dd74b67724","Type":"ContainerStarted","Data":"c1b7158217a4729312096d76c3b83c4b98bcc010a536ec92a57435302afd4fd1"} Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.317212 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.321143 4926 generic.go:334] "Generic (PLEG): container finished" podID="0ffe198c-1848-458f-988f-d8df3e03f628" containerID="8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240" exitCode=0 Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.321184 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-wffhz" event={"ID":"0ffe198c-1848-458f-988f-d8df3e03f628","Type":"ContainerDied","Data":"8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240"} Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.321220 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-869f779d85-wffhz" event={"ID":"0ffe198c-1848-458f-988f-d8df3e03f628","Type":"ContainerDied","Data":"0ad687fd03407350f3d3be9d012b503aa07d112b06529da4581db4e2c3cdbf3d"} Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.321240 4926 scope.go:117] "RemoveContainer" containerID="8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.321251 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-869f779d85-wffhz" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.367434 4926 scope.go:117] "RemoveContainer" containerID="3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.395888 4926 scope.go:117] "RemoveContainer" containerID="8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240" Nov 25 10:06:50 crc kubenswrapper[4926]: E1125 10:06:50.400334 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240\": container with ID starting with 8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240 not found: ID does not exist" containerID="8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.400376 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240"} err="failed to get container status \"8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240\": rpc error: code = NotFound desc = could not find container \"8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240\": container with ID starting with 8e6af2edf9840cbffc9f0186f4eadd8dd08c6d978b5818af4fcd80d616713240 not found: ID does not exist" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.400401 4926 scope.go:117] "RemoveContainer" containerID="3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.400545 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gskcw\" (UniqueName: \"kubernetes.io/projected/0ffe198c-1848-458f-988f-d8df3e03f628-kube-api-access-gskcw\") pod \"0ffe198c-1848-458f-988f-d8df3e03f628\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.400751 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-nb\") pod \"0ffe198c-1848-458f-988f-d8df3e03f628\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.400802 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-config\") pod \"0ffe198c-1848-458f-988f-d8df3e03f628\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.400955 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-sb\") pod \"0ffe198c-1848-458f-988f-d8df3e03f628\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.401011 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-dns-svc\") pod \"0ffe198c-1848-458f-988f-d8df3e03f628\" (UID: \"0ffe198c-1848-458f-988f-d8df3e03f628\") " Nov 25 10:06:50 crc kubenswrapper[4926]: E1125 10:06:50.402553 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53\": container with ID starting with 3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53 not found: ID does not exist" containerID="3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.402802 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53"} err="failed to get container status \"3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53\": rpc error: code = NotFound desc = could not find container \"3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53\": container with ID starting with 3c8a606e9813f788f9110d5bd57974ce39a919d31148be7e93726d34ed0c4c53 not found: ID does not exist" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.404824 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ffe198c-1848-458f-988f-d8df3e03f628-kube-api-access-gskcw" (OuterVolumeSpecName: "kube-api-access-gskcw") pod "0ffe198c-1848-458f-988f-d8df3e03f628" (UID: "0ffe198c-1848-458f-988f-d8df3e03f628"). InnerVolumeSpecName "kube-api-access-gskcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.460313 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0ffe198c-1848-458f-988f-d8df3e03f628" (UID: "0ffe198c-1848-458f-988f-d8df3e03f628"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.464665 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0ffe198c-1848-458f-988f-d8df3e03f628" (UID: "0ffe198c-1848-458f-988f-d8df3e03f628"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.474117 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0ffe198c-1848-458f-988f-d8df3e03f628" (UID: "0ffe198c-1848-458f-988f-d8df3e03f628"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.488545 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-config" (OuterVolumeSpecName: "config") pod "0ffe198c-1848-458f-988f-d8df3e03f628" (UID: "0ffe198c-1848-458f-988f-d8df3e03f628"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.502647 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.502680 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.502690 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.502697 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0ffe198c-1848-458f-988f-d8df3e03f628-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.502708 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gskcw\" (UniqueName: \"kubernetes.io/projected/0ffe198c-1848-458f-988f-d8df3e03f628-kube-api-access-gskcw\") on node \"crc\" DevicePath \"\"" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.647666 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-pz6nn"] Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.667864 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-wffhz"] Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.677320 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-869f779d85-wffhz"] Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.697763 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:06:50 crc kubenswrapper[4926]: I1125 10:06:50.831414 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:06:51 crc kubenswrapper[4926]: I1125 10:06:51.350194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0405b7fb-14c1-4958-a25a-560bfd8bf859","Type":"ContainerStarted","Data":"56c3c225ea728ccc6748fb71702ebc52180f7b9f474a77191c619ca04ac437a3"} Nov 25 10:06:51 crc kubenswrapper[4926]: I1125 10:06:51.353147 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerStarted","Data":"38a394d53d549a65130a94247267d96b77401d1e71fbe2911945a62931f9afc5"} Nov 25 10:06:51 crc kubenswrapper[4926]: I1125 10:06:51.356755 4926 generic.go:334] "Generic (PLEG): container finished" podID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerID="1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490" exitCode=0 Nov 25 10:06:51 crc kubenswrapper[4926]: I1125 10:06:51.356831 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" event={"ID":"44037126-caa3-43d2-abf4-937f10d8fe2a","Type":"ContainerDied","Data":"1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490"} Nov 25 10:06:51 crc kubenswrapper[4926]: I1125 10:06:51.356862 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" event={"ID":"44037126-caa3-43d2-abf4-937f10d8fe2a","Type":"ContainerStarted","Data":"5c0cdcd0b912d9e543b102c6c2bb1305eac4ebf5660ff67be06aba4e2a901c39"} Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.026261 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" path="/var/lib/kubelet/pods/0ffe198c-1848-458f-988f-d8df3e03f628/volumes" Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.229532 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.370243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerStarted","Data":"df00c8c47c20fcb0c08de9a847ce27e3a0f4d65adfc8cbca8609d49a2eec630e"} Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.371532 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.381452 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" event={"ID":"44037126-caa3-43d2-abf4-937f10d8fe2a","Type":"ContainerStarted","Data":"57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f"} Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.382338 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.384356 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0405b7fb-14c1-4958-a25a-560bfd8bf859","Type":"ContainerStarted","Data":"6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b"} Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.384384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0405b7fb-14c1-4958-a25a-560bfd8bf859","Type":"ContainerStarted","Data":"347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8"} Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.384492 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api-log" containerID="cri-o://347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8" gracePeriod=30 Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.384751 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.384794 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api" containerID="cri-o://6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b" gracePeriod=30 Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.442360 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.762566549 podStartE2EDuration="5.442324307s" podCreationTimestamp="2025-11-25 10:06:47 +0000 UTC" firstStartedPulling="2025-11-25 10:06:48.134094522 +0000 UTC m=+1104.818050949" lastFinishedPulling="2025-11-25 10:06:51.81385228 +0000 UTC m=+1108.497808707" observedRunningTime="2025-11-25 10:06:52.434336928 +0000 UTC m=+1109.118293355" watchObservedRunningTime="2025-11-25 10:06:52.442324307 +0000 UTC m=+1109.126280734" Nov 25 10:06:52 crc kubenswrapper[4926]: I1125 10:06:52.534046 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" podStartSLOduration=3.534023223 podStartE2EDuration="3.534023223s" podCreationTimestamp="2025-11-25 10:06:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:52.491876862 +0000 UTC m=+1109.175833289" watchObservedRunningTime="2025-11-25 10:06:52.534023223 +0000 UTC m=+1109.217979650" Nov 25 10:06:53 crc kubenswrapper[4926]: I1125 10:06:53.420216 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ca6a1d2-b396-44e8-9701-f3dd74b67724","Type":"ContainerStarted","Data":"ccc91f9557a6fa6652437d47b03a6621892f4960ab6c510d107e7ff37cb15be1"} Nov 25 10:06:53 crc kubenswrapper[4926]: I1125 10:06:53.422629 4926 generic.go:334] "Generic (PLEG): container finished" podID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerID="347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8" exitCode=143 Nov 25 10:06:53 crc kubenswrapper[4926]: I1125 10:06:53.422689 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0405b7fb-14c1-4958-a25a-560bfd8bf859","Type":"ContainerDied","Data":"347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8"} Nov 25 10:06:53 crc kubenswrapper[4926]: I1125 10:06:53.521979 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:06:53 crc kubenswrapper[4926]: I1125 10:06:53.547778 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.547755493 podStartE2EDuration="4.547755493s" podCreationTimestamp="2025-11-25 10:06:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:06:52.538421773 +0000 UTC m=+1109.222378200" watchObservedRunningTime="2025-11-25 10:06:53.547755493 +0000 UTC m=+1110.231711920" Nov 25 10:06:54 crc kubenswrapper[4926]: I1125 10:06:54.454080 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ca6a1d2-b396-44e8-9701-f3dd74b67724","Type":"ContainerStarted","Data":"3e7ffe4eeba92a827a2c4e7a4f83b132f2e40ec8f687f579f1b552518b527f43"} Nov 25 10:06:54 crc kubenswrapper[4926]: I1125 10:06:54.479361 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.031994457 podStartE2EDuration="5.479345015s" podCreationTimestamp="2025-11-25 10:06:49 +0000 UTC" firstStartedPulling="2025-11-25 10:06:50.26781022 +0000 UTC m=+1106.951766657" lastFinishedPulling="2025-11-25 10:06:52.715160788 +0000 UTC m=+1109.399117215" observedRunningTime="2025-11-25 10:06:54.478519835 +0000 UTC m=+1111.162476262" watchObservedRunningTime="2025-11-25 10:06:54.479345015 +0000 UTC m=+1111.163301442" Nov 25 10:06:54 crc kubenswrapper[4926]: I1125 10:06:54.882868 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 10:06:55 crc kubenswrapper[4926]: I1125 10:06:55.931093 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6cdc678f6f-htchg" Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.006833 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76b9fdf556-hhq5f"] Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.007484 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76b9fdf556-hhq5f" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-api" containerID="cri-o://03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0" gracePeriod=30 Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.008169 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-76b9fdf556-hhq5f" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-httpd" containerID="cri-o://b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a" gracePeriod=30 Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.379060 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.473678 4926 generic.go:334] "Generic (PLEG): container finished" podID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerID="b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a" exitCode=0 Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.474602 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b9fdf556-hhq5f" event={"ID":"408f4021-7f18-46a6-b6e8-ef300c2f176a","Type":"ContainerDied","Data":"b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a"} Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.507971 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-8594cdbfd6-wcgzz" Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.601098 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58855fbcfd-9r5c5"] Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.601325 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58855fbcfd-9r5c5" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api-log" containerID="cri-o://4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101" gracePeriod=30 Nov 25 10:06:56 crc kubenswrapper[4926]: I1125 10:06:56.601690 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-58855fbcfd-9r5c5" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api" containerID="cri-o://a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94" gracePeriod=30 Nov 25 10:06:57 crc kubenswrapper[4926]: I1125 10:06:57.482704 4926 generic.go:334] "Generic (PLEG): container finished" podID="9083ac42-380c-471a-a89f-27c6318e5493" containerID="4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101" exitCode=143 Nov 25 10:06:57 crc kubenswrapper[4926]: I1125 10:06:57.482745 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58855fbcfd-9r5c5" event={"ID":"9083ac42-380c-471a-a89f-27c6318e5493","Type":"ContainerDied","Data":"4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101"} Nov 25 10:06:57 crc kubenswrapper[4926]: I1125 10:06:57.648652 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:57 crc kubenswrapper[4926]: I1125 10:06:57.687768 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-d785fcd8-5cqwz" Nov 25 10:06:59 crc kubenswrapper[4926]: I1125 10:06:59.754925 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-58855fbcfd-9r5c5" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": read tcp 10.217.0.2:48424->10.217.0.147:9311: read: connection reset by peer" Nov 25 10:06:59 crc kubenswrapper[4926]: I1125 10:06:59.755014 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-58855fbcfd-9r5c5" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.147:9311/healthcheck\": read tcp 10.217.0.2:48434->10.217.0.147:9311: read: connection reset by peer" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.061167 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.109828 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.129042 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-r7r2g"] Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.129308 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerName="dnsmasq-dns" containerID="cri-o://9f5e0deb21fa04290066b2c4d6c914c6a47f0dfd82e4a59004cde764f25609a2" gracePeriod=10 Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.190268 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.451040 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.464328 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.508778 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data\") pod \"9083ac42-380c-471a-a89f-27c6318e5493\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.508881 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5jng\" (UniqueName: \"kubernetes.io/projected/9083ac42-380c-471a-a89f-27c6318e5493-kube-api-access-l5jng\") pod \"9083ac42-380c-471a-a89f-27c6318e5493\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.508907 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-combined-ca-bundle\") pod \"9083ac42-380c-471a-a89f-27c6318e5493\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.509007 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data-custom\") pod \"9083ac42-380c-471a-a89f-27c6318e5493\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.509045 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9083ac42-380c-471a-a89f-27c6318e5493-logs\") pod \"9083ac42-380c-471a-a89f-27c6318e5493\" (UID: \"9083ac42-380c-471a-a89f-27c6318e5493\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.510047 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9083ac42-380c-471a-a89f-27c6318e5493-logs" (OuterVolumeSpecName: "logs") pod "9083ac42-380c-471a-a89f-27c6318e5493" (UID: "9083ac42-380c-471a-a89f-27c6318e5493"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.521464 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9083ac42-380c-471a-a89f-27c6318e5493-kube-api-access-l5jng" (OuterVolumeSpecName: "kube-api-access-l5jng") pod "9083ac42-380c-471a-a89f-27c6318e5493" (UID: "9083ac42-380c-471a-a89f-27c6318e5493"). InnerVolumeSpecName "kube-api-access-l5jng". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.522986 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9083ac42-380c-471a-a89f-27c6318e5493" (UID: "9083ac42-380c-471a-a89f-27c6318e5493"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.530272 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" event={"ID":"14ff3b9a-634a-433c-a454-3d09f6fddbf0","Type":"ContainerDied","Data":"9f5e0deb21fa04290066b2c4d6c914c6a47f0dfd82e4a59004cde764f25609a2"} Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.530345 4926 generic.go:334] "Generic (PLEG): container finished" podID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerID="9f5e0deb21fa04290066b2c4d6c914c6a47f0dfd82e4a59004cde764f25609a2" exitCode=0 Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.540348 4926 generic.go:334] "Generic (PLEG): container finished" podID="9083ac42-380c-471a-a89f-27c6318e5493" containerID="a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94" exitCode=0 Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.540462 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58855fbcfd-9r5c5" event={"ID":"9083ac42-380c-471a-a89f-27c6318e5493","Type":"ContainerDied","Data":"a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94"} Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.540526 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-58855fbcfd-9r5c5" event={"ID":"9083ac42-380c-471a-a89f-27c6318e5493","Type":"ContainerDied","Data":"6ce0568bf33422a9fff2c9de5a457ae044666be6c20c889a6b7618ef2f2ff3ae"} Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.540549 4926 scope.go:117] "RemoveContainer" containerID="a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.541139 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-58855fbcfd-9r5c5" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.545907 4926 generic.go:334] "Generic (PLEG): container finished" podID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerID="03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0" exitCode=0 Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.546071 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76b9fdf556-hhq5f" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.546116 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="cinder-scheduler" containerID="cri-o://ccc91f9557a6fa6652437d47b03a6621892f4960ab6c510d107e7ff37cb15be1" gracePeriod=30 Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.546342 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b9fdf556-hhq5f" event={"ID":"408f4021-7f18-46a6-b6e8-ef300c2f176a","Type":"ContainerDied","Data":"03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0"} Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.546374 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76b9fdf556-hhq5f" event={"ID":"408f4021-7f18-46a6-b6e8-ef300c2f176a","Type":"ContainerDied","Data":"1665be9afb3ead929a0e6c15650895bd2f659807b3001fc797c875f258dd28ac"} Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.546648 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="probe" containerID="cri-o://3e7ffe4eeba92a827a2c4e7a4f83b132f2e40ec8f687f579f1b552518b527f43" gracePeriod=30 Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.554496 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9083ac42-380c-471a-a89f-27c6318e5493" (UID: "9083ac42-380c-471a-a89f-27c6318e5493"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.610885 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-httpd-config\") pod \"408f4021-7f18-46a6-b6e8-ef300c2f176a\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611046 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-ovndb-tls-certs\") pod \"408f4021-7f18-46a6-b6e8-ef300c2f176a\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611094 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-combined-ca-bundle\") pod \"408f4021-7f18-46a6-b6e8-ef300c2f176a\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611166 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-config\") pod \"408f4021-7f18-46a6-b6e8-ef300c2f176a\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611292 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffq7d\" (UniqueName: \"kubernetes.io/projected/408f4021-7f18-46a6-b6e8-ef300c2f176a-kube-api-access-ffq7d\") pod \"408f4021-7f18-46a6-b6e8-ef300c2f176a\" (UID: \"408f4021-7f18-46a6-b6e8-ef300c2f176a\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611909 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5jng\" (UniqueName: \"kubernetes.io/projected/9083ac42-380c-471a-a89f-27c6318e5493-kube-api-access-l5jng\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611955 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611968 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.611980 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9083ac42-380c-471a-a89f-27c6318e5493-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.618091 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "408f4021-7f18-46a6-b6e8-ef300c2f176a" (UID: "408f4021-7f18-46a6-b6e8-ef300c2f176a"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.621675 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/408f4021-7f18-46a6-b6e8-ef300c2f176a-kube-api-access-ffq7d" (OuterVolumeSpecName: "kube-api-access-ffq7d") pod "408f4021-7f18-46a6-b6e8-ef300c2f176a" (UID: "408f4021-7f18-46a6-b6e8-ef300c2f176a"). InnerVolumeSpecName "kube-api-access-ffq7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.634112 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data" (OuterVolumeSpecName: "config-data") pod "9083ac42-380c-471a-a89f-27c6318e5493" (UID: "9083ac42-380c-471a-a89f-27c6318e5493"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.676448 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-config" (OuterVolumeSpecName: "config") pod "408f4021-7f18-46a6-b6e8-ef300c2f176a" (UID: "408f4021-7f18-46a6-b6e8-ef300c2f176a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.689506 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "408f4021-7f18-46a6-b6e8-ef300c2f176a" (UID: "408f4021-7f18-46a6-b6e8-ef300c2f176a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.697751 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "408f4021-7f18-46a6-b6e8-ef300c2f176a" (UID: "408f4021-7f18-46a6-b6e8-ef300c2f176a"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.717063 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffq7d\" (UniqueName: \"kubernetes.io/projected/408f4021-7f18-46a6-b6e8-ef300c2f176a-kube-api-access-ffq7d\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.717109 4926 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-httpd-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.717125 4926 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.717139 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.717152 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/408f4021-7f18-46a6-b6e8-ef300c2f176a-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.717165 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9083ac42-380c-471a-a89f-27c6318e5493-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.757094 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.767397 4926 scope.go:117] "RemoveContainer" containerID="4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.795651 4926 scope.go:117] "RemoveContainer" containerID="a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94" Nov 25 10:07:00 crc kubenswrapper[4926]: E1125 10:07:00.796264 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94\": container with ID starting with a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94 not found: ID does not exist" containerID="a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.796357 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94"} err="failed to get container status \"a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94\": rpc error: code = NotFound desc = could not find container \"a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94\": container with ID starting with a8d21b3bd0304067ebae25f53839c305f40f1b7026ff35a89388ddae6b02bf94 not found: ID does not exist" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.796446 4926 scope.go:117] "RemoveContainer" containerID="4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101" Nov 25 10:07:00 crc kubenswrapper[4926]: E1125 10:07:00.796785 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101\": container with ID starting with 4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101 not found: ID does not exist" containerID="4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.796835 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101"} err="failed to get container status \"4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101\": rpc error: code = NotFound desc = could not find container \"4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101\": container with ID starting with 4db3a46991b57b949b3f29ac00b3113b47e0a424bf0a5db763b85880f9005101 not found: ID does not exist" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.796868 4926 scope.go:117] "RemoveContainer" containerID="b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.815441 4926 scope.go:117] "RemoveContainer" containerID="03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.822424 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-dns-svc\") pod \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.822565 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-sb\") pod \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.822664 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-config\") pod \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.831236 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-nb\") pod \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.831399 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vstw\" (UniqueName: \"kubernetes.io/projected/14ff3b9a-634a-433c-a454-3d09f6fddbf0-kube-api-access-8vstw\") pod \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\" (UID: \"14ff3b9a-634a-433c-a454-3d09f6fddbf0\") " Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.837634 4926 scope.go:117] "RemoveContainer" containerID="b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.838455 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/14ff3b9a-634a-433c-a454-3d09f6fddbf0-kube-api-access-8vstw" (OuterVolumeSpecName: "kube-api-access-8vstw") pod "14ff3b9a-634a-433c-a454-3d09f6fddbf0" (UID: "14ff3b9a-634a-433c-a454-3d09f6fddbf0"). InnerVolumeSpecName "kube-api-access-8vstw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: E1125 10:07:00.839270 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a\": container with ID starting with b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a not found: ID does not exist" containerID="b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.839320 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a"} err="failed to get container status \"b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a\": rpc error: code = NotFound desc = could not find container \"b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a\": container with ID starting with b99f80d602ab63a8f2017a0d4e53e6fc27b891223e09670621e198bdc22c479a not found: ID does not exist" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.839343 4926 scope.go:117] "RemoveContainer" containerID="03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0" Nov 25 10:07:00 crc kubenswrapper[4926]: E1125 10:07:00.839669 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0\": container with ID starting with 03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0 not found: ID does not exist" containerID="03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.839714 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0"} err="failed to get container status \"03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0\": rpc error: code = NotFound desc = could not find container \"03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0\": container with ID starting with 03e590ed39889331a8a4bc6709231b33eec5bc0e1a0b1b016cf327a01cfd30b0 not found: ID does not exist" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.887541 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-config" (OuterVolumeSpecName: "config") pod "14ff3b9a-634a-433c-a454-3d09f6fddbf0" (UID: "14ff3b9a-634a-433c-a454-3d09f6fddbf0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.896928 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "14ff3b9a-634a-433c-a454-3d09f6fddbf0" (UID: "14ff3b9a-634a-433c-a454-3d09f6fddbf0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.900427 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-76b9fdf556-hhq5f"] Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.901422 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "14ff3b9a-634a-433c-a454-3d09f6fddbf0" (UID: "14ff3b9a-634a-433c-a454-3d09f6fddbf0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.909464 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "14ff3b9a-634a-433c-a454-3d09f6fddbf0" (UID: "14ff3b9a-634a-433c-a454-3d09f6fddbf0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.909855 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-76b9fdf556-hhq5f"] Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.918193 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-58855fbcfd-9r5c5"] Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.924601 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-58855fbcfd-9r5c5"] Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.939248 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.939281 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.939293 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.939302 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/14ff3b9a-634a-433c-a454-3d09f6fddbf0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:00 crc kubenswrapper[4926]: I1125 10:07:00.939315 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vstw\" (UniqueName: \"kubernetes.io/projected/14ff3b9a-634a-433c-a454-3d09f6fddbf0-kube-api-access-8vstw\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.557469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" event={"ID":"14ff3b9a-634a-433c-a454-3d09f6fddbf0","Type":"ContainerDied","Data":"8cd32010ad7a8cc16d6b59f4b89ace02aac3a03b2b70b39987378df5fdef26dd"} Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.557505 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f66db59b9-r7r2g" Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.557879 4926 scope.go:117] "RemoveContainer" containerID="9f5e0deb21fa04290066b2c4d6c914c6a47f0dfd82e4a59004cde764f25609a2" Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.569112 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerID="3e7ffe4eeba92a827a2c4e7a4f83b132f2e40ec8f687f579f1b552518b527f43" exitCode=0 Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.569166 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ca6a1d2-b396-44e8-9701-f3dd74b67724","Type":"ContainerDied","Data":"3e7ffe4eeba92a827a2c4e7a4f83b132f2e40ec8f687f579f1b552518b527f43"} Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.618142 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-r7r2g"] Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.620426 4926 scope.go:117] "RemoveContainer" containerID="7aa1f95a630fa036b347fb8ffd5d3f1b67a018db89a6b91ecb06548144f58364" Nov 25 10:07:01 crc kubenswrapper[4926]: I1125 10:07:01.628116 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f66db59b9-r7r2g"] Nov 25 10:07:02 crc kubenswrapper[4926]: I1125 10:07:02.024176 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" path="/var/lib/kubelet/pods/14ff3b9a-634a-433c-a454-3d09f6fddbf0/volumes" Nov 25 10:07:02 crc kubenswrapper[4926]: I1125 10:07:02.024987 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" path="/var/lib/kubelet/pods/408f4021-7f18-46a6-b6e8-ef300c2f176a/volumes" Nov 25 10:07:02 crc kubenswrapper[4926]: I1125 10:07:02.025804 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9083ac42-380c-471a-a89f-27c6318e5493" path="/var/lib/kubelet/pods/9083ac42-380c-471a-a89f-27c6318e5493/volumes" Nov 25 10:07:02 crc kubenswrapper[4926]: I1125 10:07:02.308004 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:07:02 crc kubenswrapper[4926]: I1125 10:07:02.441145 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 10:07:03 crc kubenswrapper[4926]: I1125 10:07:03.889635 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-5c5477c6d9-t4dvb" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.114589 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.115171 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" containerName="dnsmasq-dns" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.115235 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" containerName="dnsmasq-dns" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.115319 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerName="dnsmasq-dns" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.115378 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerName="dnsmasq-dns" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.115426 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerName="init" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.115474 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerName="init" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.117401 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.117461 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.117516 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-httpd" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.117577 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-httpd" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.117637 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api-log" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.117683 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api-log" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.117917 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-api" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.117985 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-api" Nov 25 10:07:04 crc kubenswrapper[4926]: E1125 10:07:04.118042 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" containerName="init" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118149 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" containerName="init" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118450 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="14ff3b9a-634a-433c-a454-3d09f6fddbf0" containerName="dnsmasq-dns" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118512 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-httpd" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118567 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api-log" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118638 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="408f4021-7f18-46a6-b6e8-ef300c2f176a" containerName="neutron-api" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118695 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ffe198c-1848-458f-988f-d8df3e03f628" containerName="dnsmasq-dns" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.118753 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="9083ac42-380c-471a-a89f-27c6318e5493" containerName="barbican-api" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.119419 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.121927 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.122174 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.122414 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-9rsbs" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.123200 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.198261 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02f6bba3-15f5-40ad-9624-deb63a912775-openstack-config\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.198300 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02f6bba3-15f5-40ad-9624-deb63a912775-openstack-config-secret\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.198448 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6bba3-15f5-40ad-9624-deb63a912775-combined-ca-bundle\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.198486 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pr47\" (UniqueName: \"kubernetes.io/projected/02f6bba3-15f5-40ad-9624-deb63a912775-kube-api-access-6pr47\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.300046 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6bba3-15f5-40ad-9624-deb63a912775-combined-ca-bundle\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.300434 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pr47\" (UniqueName: \"kubernetes.io/projected/02f6bba3-15f5-40ad-9624-deb63a912775-kube-api-access-6pr47\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.300517 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02f6bba3-15f5-40ad-9624-deb63a912775-openstack-config\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.300543 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02f6bba3-15f5-40ad-9624-deb63a912775-openstack-config-secret\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.301411 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/02f6bba3-15f5-40ad-9624-deb63a912775-openstack-config\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.316343 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02f6bba3-15f5-40ad-9624-deb63a912775-combined-ca-bundle\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.316451 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/02f6bba3-15f5-40ad-9624-deb63a912775-openstack-config-secret\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.320208 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pr47\" (UniqueName: \"kubernetes.io/projected/02f6bba3-15f5-40ad-9624-deb63a912775-kube-api-access-6pr47\") pod \"openstackclient\" (UID: \"02f6bba3-15f5-40ad-9624-deb63a912775\") " pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.449886 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.614110 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerID="ccc91f9557a6fa6652437d47b03a6621892f4960ab6c510d107e7ff37cb15be1" exitCode=0 Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.614228 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ca6a1d2-b396-44e8-9701-f3dd74b67724","Type":"ContainerDied","Data":"ccc91f9557a6fa6652437d47b03a6621892f4960ab6c510d107e7ff37cb15be1"} Nov 25 10:07:04 crc kubenswrapper[4926]: I1125 10:07:04.910625 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020463 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data\") pod \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020505 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data-custom\") pod \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020636 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-scripts\") pod \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020721 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-combined-ca-bundle\") pod \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ca6a1d2-b396-44e8-9701-f3dd74b67724-etc-machine-id\") pod \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020772 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jkw4\" (UniqueName: \"kubernetes.io/projected/1ca6a1d2-b396-44e8-9701-f3dd74b67724-kube-api-access-4jkw4\") pod \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\" (UID: \"1ca6a1d2-b396-44e8-9701-f3dd74b67724\") " Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.020863 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ca6a1d2-b396-44e8-9701-f3dd74b67724-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1ca6a1d2-b396-44e8-9701-f3dd74b67724" (UID: "1ca6a1d2-b396-44e8-9701-f3dd74b67724"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.021145 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ca6a1d2-b396-44e8-9701-f3dd74b67724-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.024870 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-scripts" (OuterVolumeSpecName: "scripts") pod "1ca6a1d2-b396-44e8-9701-f3dd74b67724" (UID: "1ca6a1d2-b396-44e8-9701-f3dd74b67724"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.025192 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "1ca6a1d2-b396-44e8-9701-f3dd74b67724" (UID: "1ca6a1d2-b396-44e8-9701-f3dd74b67724"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.026466 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ca6a1d2-b396-44e8-9701-f3dd74b67724-kube-api-access-4jkw4" (OuterVolumeSpecName: "kube-api-access-4jkw4") pod "1ca6a1d2-b396-44e8-9701-f3dd74b67724" (UID: "1ca6a1d2-b396-44e8-9701-f3dd74b67724"). InnerVolumeSpecName "kube-api-access-4jkw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.055365 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.078965 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1ca6a1d2-b396-44e8-9701-f3dd74b67724" (UID: "1ca6a1d2-b396-44e8-9701-f3dd74b67724"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.122927 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.122980 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.122996 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jkw4\" (UniqueName: \"kubernetes.io/projected/1ca6a1d2-b396-44e8-9701-f3dd74b67724-kube-api-access-4jkw4\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.123011 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.127611 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data" (OuterVolumeSpecName: "config-data") pod "1ca6a1d2-b396-44e8-9701-f3dd74b67724" (UID: "1ca6a1d2-b396-44e8-9701-f3dd74b67724"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.224862 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ca6a1d2-b396-44e8-9701-f3dd74b67724-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.625714 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"1ca6a1d2-b396-44e8-9701-f3dd74b67724","Type":"ContainerDied","Data":"c1b7158217a4729312096d76c3b83c4b98bcc010a536ec92a57435302afd4fd1"} Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.625749 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.626019 4926 scope.go:117] "RemoveContainer" containerID="3e7ffe4eeba92a827a2c4e7a4f83b132f2e40ec8f687f579f1b552518b527f43" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.630560 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"02f6bba3-15f5-40ad-9624-deb63a912775","Type":"ContainerStarted","Data":"6b0110c7d9c40b0375d00dc7b397b2f204e999b69b4d2559ad4379f5f6aace87"} Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.662169 4926 scope.go:117] "RemoveContainer" containerID="ccc91f9557a6fa6652437d47b03a6621892f4960ab6c510d107e7ff37cb15be1" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.676143 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.688534 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.700365 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:07:05 crc kubenswrapper[4926]: E1125 10:07:05.700734 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="cinder-scheduler" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.700750 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="cinder-scheduler" Nov 25 10:07:05 crc kubenswrapper[4926]: E1125 10:07:05.700776 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="probe" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.700782 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="probe" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.700962 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="cinder-scheduler" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.700999 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" containerName="probe" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.712175 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.715812 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.723539 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.837444 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.837513 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-scripts\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.837596 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.837643 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-config-data\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.837665 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d2781eb-b45c-4445-9f30-009ab5de333f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.837772 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbb6d\" (UniqueName: \"kubernetes.io/projected/5d2781eb-b45c-4445-9f30-009ab5de333f-kube-api-access-sbb6d\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.939129 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.939197 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-scripts\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.939261 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.939305 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-config-data\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.939332 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d2781eb-b45c-4445-9f30-009ab5de333f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.939394 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbb6d\" (UniqueName: \"kubernetes.io/projected/5d2781eb-b45c-4445-9f30-009ab5de333f-kube-api-access-sbb6d\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.940828 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5d2781eb-b45c-4445-9f30-009ab5de333f-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.945760 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.945986 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-config-data\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.949349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-scripts\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.949439 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d2781eb-b45c-4445-9f30-009ab5de333f-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:05 crc kubenswrapper[4926]: I1125 10:07:05.959363 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbb6d\" (UniqueName: \"kubernetes.io/projected/5d2781eb-b45c-4445-9f30-009ab5de333f-kube-api-access-sbb6d\") pod \"cinder-scheduler-0\" (UID: \"5d2781eb-b45c-4445-9f30-009ab5de333f\") " pod="openstack/cinder-scheduler-0" Nov 25 10:07:06 crc kubenswrapper[4926]: I1125 10:07:06.028393 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ca6a1d2-b396-44e8-9701-f3dd74b67724" path="/var/lib/kubelet/pods/1ca6a1d2-b396-44e8-9701-f3dd74b67724/volumes" Nov 25 10:07:06 crc kubenswrapper[4926]: I1125 10:07:06.058297 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Nov 25 10:07:06 crc kubenswrapper[4926]: I1125 10:07:06.601902 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Nov 25 10:07:06 crc kubenswrapper[4926]: I1125 10:07:06.646067 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5d2781eb-b45c-4445-9f30-009ab5de333f","Type":"ContainerStarted","Data":"9c2e2b331c461a73b4c4f7cf4944405f76d7a0cb48539a3d611daa3eec0b52df"} Nov 25 10:07:07 crc kubenswrapper[4926]: I1125 10:07:07.657537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5d2781eb-b45c-4445-9f30-009ab5de333f","Type":"ContainerStarted","Data":"c36ea049a2fd9f616892ae78f4758b476ccfeeaabb67a99c545f314441e14254"} Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.290294 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pr557"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.291898 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.307573 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pr557"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.381091 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-jqn4l"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.382883 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.388529 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-operator-scripts\") pod \"nova-api-db-create-pr557\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.388749 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rsd5\" (UniqueName: \"kubernetes.io/projected/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-kube-api-access-5rsd5\") pod \"nova-api-db-create-pr557\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.406464 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jqn4l"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.490560 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rsd5\" (UniqueName: \"kubernetes.io/projected/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-kube-api-access-5rsd5\") pod \"nova-api-db-create-pr557\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.490635 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-operator-scripts\") pod \"nova-cell0-db-create-jqn4l\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.490681 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-operator-scripts\") pod \"nova-api-db-create-pr557\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.490711 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sckk\" (UniqueName: \"kubernetes.io/projected/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-kube-api-access-7sckk\") pod \"nova-cell0-db-create-jqn4l\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.491566 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-operator-scripts\") pod \"nova-api-db-create-pr557\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.501872 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0b02-account-create-jpmcd"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.503026 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.506594 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.511197 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0b02-account-create-jpmcd"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.528394 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rsd5\" (UniqueName: \"kubernetes.io/projected/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-kube-api-access-5rsd5\") pod \"nova-api-db-create-pr557\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.582813 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-w4qxc"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.584311 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.593174 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d24462-3c87-4454-bf92-e97345e32e12-operator-scripts\") pod \"nova-api-0b02-account-create-jpmcd\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.593453 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rtc7\" (UniqueName: \"kubernetes.io/projected/c7d24462-3c87-4454-bf92-e97345e32e12-kube-api-access-5rtc7\") pod \"nova-api-0b02-account-create-jpmcd\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.593489 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-operator-scripts\") pod \"nova-cell0-db-create-jqn4l\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.593529 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sckk\" (UniqueName: \"kubernetes.io/projected/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-kube-api-access-7sckk\") pod \"nova-cell0-db-create-jqn4l\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.594407 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-operator-scripts\") pod \"nova-cell0-db-create-jqn4l\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.613989 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-w4qxc"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.619269 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.642664 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sckk\" (UniqueName: \"kubernetes.io/projected/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-kube-api-access-7sckk\") pod \"nova-cell0-db-create-jqn4l\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.696171 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5d2781eb-b45c-4445-9f30-009ab5de333f","Type":"ContainerStarted","Data":"cef6d08cec4e5b47977b498a70fc81da797ca363c7c32e4d034981fe97a2590c"} Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.697044 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pjzv\" (UniqueName: \"kubernetes.io/projected/c58cf150-5525-4d1c-9376-a3bd78100269-kube-api-access-4pjzv\") pod \"nova-cell1-db-create-w4qxc\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.697106 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c58cf150-5525-4d1c-9376-a3bd78100269-operator-scripts\") pod \"nova-cell1-db-create-w4qxc\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.697180 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d24462-3c87-4454-bf92-e97345e32e12-operator-scripts\") pod \"nova-api-0b02-account-create-jpmcd\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.697209 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rtc7\" (UniqueName: \"kubernetes.io/projected/c7d24462-3c87-4454-bf92-e97345e32e12-kube-api-access-5rtc7\") pod \"nova-api-0b02-account-create-jpmcd\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.697358 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f822-account-create-hghgh"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.698006 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d24462-3c87-4454-bf92-e97345e32e12-operator-scripts\") pod \"nova-api-0b02-account-create-jpmcd\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.698814 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.701614 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.704039 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.722519 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f822-account-create-hghgh"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.729243 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rtc7\" (UniqueName: \"kubernetes.io/projected/c7d24462-3c87-4454-bf92-e97345e32e12-kube-api-access-5rtc7\") pod \"nova-api-0b02-account-create-jpmcd\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.736430 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.7364085830000002 podStartE2EDuration="3.736408583s" podCreationTimestamp="2025-11-25 10:07:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:07:08.729719487 +0000 UTC m=+1125.413675914" watchObservedRunningTime="2025-11-25 10:07:08.736408583 +0000 UTC m=+1125.420365000" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.799472 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81121ebf-1d21-43de-8787-efc9b79578e2-operator-scripts\") pod \"nova-cell0-f822-account-create-hghgh\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.799569 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pjzv\" (UniqueName: \"kubernetes.io/projected/c58cf150-5525-4d1c-9376-a3bd78100269-kube-api-access-4pjzv\") pod \"nova-cell1-db-create-w4qxc\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.799619 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdb7g\" (UniqueName: \"kubernetes.io/projected/81121ebf-1d21-43de-8787-efc9b79578e2-kube-api-access-hdb7g\") pod \"nova-cell0-f822-account-create-hghgh\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.799657 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c58cf150-5525-4d1c-9376-a3bd78100269-operator-scripts\") pod \"nova-cell1-db-create-w4qxc\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.803314 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c58cf150-5525-4d1c-9376-a3bd78100269-operator-scripts\") pod \"nova-cell1-db-create-w4qxc\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.817008 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.829110 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pjzv\" (UniqueName: \"kubernetes.io/projected/c58cf150-5525-4d1c-9376-a3bd78100269-kube-api-access-4pjzv\") pod \"nova-cell1-db-create-w4qxc\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.898146 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-dded-account-create-scxpg"] Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.899383 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.901008 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81121ebf-1d21-43de-8787-efc9b79578e2-operator-scripts\") pod \"nova-cell0-f822-account-create-hghgh\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.901071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdb7g\" (UniqueName: \"kubernetes.io/projected/81121ebf-1d21-43de-8787-efc9b79578e2-kube-api-access-hdb7g\") pod \"nova-cell0-f822-account-create-hghgh\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.901401 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.901865 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81121ebf-1d21-43de-8787-efc9b79578e2-operator-scripts\") pod \"nova-cell0-f822-account-create-hghgh\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.902639 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.919644 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdb7g\" (UniqueName: \"kubernetes.io/projected/81121ebf-1d21-43de-8787-efc9b79578e2-kube-api-access-hdb7g\") pod \"nova-cell0-f822-account-create-hghgh\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:08 crc kubenswrapper[4926]: I1125 10:07:08.925730 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-dded-account-create-scxpg"] Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.003876 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7b372d-b0e0-4423-9734-00d223b0e99e-operator-scripts\") pod \"nova-cell1-dded-account-create-scxpg\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.004008 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdhb\" (UniqueName: \"kubernetes.io/projected/2e7b372d-b0e0-4423-9734-00d223b0e99e-kube-api-access-lkdhb\") pod \"nova-cell1-dded-account-create-scxpg\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.105756 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7b372d-b0e0-4423-9734-00d223b0e99e-operator-scripts\") pod \"nova-cell1-dded-account-create-scxpg\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.105903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdhb\" (UniqueName: \"kubernetes.io/projected/2e7b372d-b0e0-4423-9734-00d223b0e99e-kube-api-access-lkdhb\") pod \"nova-cell1-dded-account-create-scxpg\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.110452 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7b372d-b0e0-4423-9734-00d223b0e99e-operator-scripts\") pod \"nova-cell1-dded-account-create-scxpg\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.136579 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.136726 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdhb\" (UniqueName: \"kubernetes.io/projected/2e7b372d-b0e0-4423-9734-00d223b0e99e-kube-api-access-lkdhb\") pod \"nova-cell1-dded-account-create-scxpg\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.146500 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pr557"] Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.229331 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.290709 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-jqn4l"] Nov 25 10:07:09 crc kubenswrapper[4926]: W1125 10:07:09.352838 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e32ccd7_0d61_4328_a22b_4637bd3f16f4.slice/crio-f6b1bd10fe9d54d9044a12274362de00201a0cd2b2de963a26682b3c52b108f6 WatchSource:0}: Error finding container f6b1bd10fe9d54d9044a12274362de00201a0cd2b2de963a26682b3c52b108f6: Status 404 returned error can't find the container with id f6b1bd10fe9d54d9044a12274362de00201a0cd2b2de963a26682b3c52b108f6 Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.381398 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0b02-account-create-jpmcd"] Nov 25 10:07:09 crc kubenswrapper[4926]: W1125 10:07:09.386386 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc7d24462_3c87_4454_bf92_e97345e32e12.slice/crio-99e2b9640d8308643197259e479d8eb156645bea857f5f025242944ad8d5eb94 WatchSource:0}: Error finding container 99e2b9640d8308643197259e479d8eb156645bea857f5f025242944ad8d5eb94: Status 404 returned error can't find the container with id 99e2b9640d8308643197259e479d8eb156645bea857f5f025242944ad8d5eb94 Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.477122 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-w4qxc"] Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.625698 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f822-account-create-hghgh"] Nov 25 10:07:09 crc kubenswrapper[4926]: W1125 10:07:09.699686 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81121ebf_1d21_43de_8787_efc9b79578e2.slice/crio-29590fd09d2ae208f2fb05be35d7c77567ea59a55c0e98a88884c8494b6d25cb WatchSource:0}: Error finding container 29590fd09d2ae208f2fb05be35d7c77567ea59a55c0e98a88884c8494b6d25cb: Status 404 returned error can't find the container with id 29590fd09d2ae208f2fb05be35d7c77567ea59a55c0e98a88884c8494b6d25cb Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.716185 4926 generic.go:334] "Generic (PLEG): container finished" podID="1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" containerID="35db8bbb0f786e7c87bad0f7902281aed6b6f6ab4b9e41939aac0ac33819b745" exitCode=0 Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.716514 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pr557" event={"ID":"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc","Type":"ContainerDied","Data":"35db8bbb0f786e7c87bad0f7902281aed6b6f6ab4b9e41939aac0ac33819b745"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.718950 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pr557" event={"ID":"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc","Type":"ContainerStarted","Data":"6cde52f52c088df59fd08dbb049d7106917bfde73d816690277e0ad6f85762a6"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.721676 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jqn4l" event={"ID":"8e32ccd7-0d61-4328-a22b-4637bd3f16f4","Type":"ContainerStarted","Data":"893aa06815f5c11ee726bd49637d5a109d184e83b517cc17a67ea83e5e444eb5"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.721711 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jqn4l" event={"ID":"8e32ccd7-0d61-4328-a22b-4637bd3f16f4","Type":"ContainerStarted","Data":"f6b1bd10fe9d54d9044a12274362de00201a0cd2b2de963a26682b3c52b108f6"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.724480 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0b02-account-create-jpmcd" event={"ID":"c7d24462-3c87-4454-bf92-e97345e32e12","Type":"ContainerStarted","Data":"e01be87fd1511b3fab3ca6d93a3ddcc5e8dee7080df81c037dfd317c9d2e6750"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.724503 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0b02-account-create-jpmcd" event={"ID":"c7d24462-3c87-4454-bf92-e97345e32e12","Type":"ContainerStarted","Data":"99e2b9640d8308643197259e479d8eb156645bea857f5f025242944ad8d5eb94"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.733498 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w4qxc" event={"ID":"c58cf150-5525-4d1c-9376-a3bd78100269","Type":"ContainerStarted","Data":"05712ab1d79bb1d97c2ac750e2ceedf1aa5bbcfe700684b4eef7b254d0ec6b8d"} Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.760370 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-jqn4l" podStartSLOduration=1.760348019 podStartE2EDuration="1.760348019s" podCreationTimestamp="2025-11-25 10:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:07:09.750272907 +0000 UTC m=+1126.434229344" watchObservedRunningTime="2025-11-25 10:07:09.760348019 +0000 UTC m=+1126.444304446" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.768121 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0b02-account-create-jpmcd" podStartSLOduration=1.766624455 podStartE2EDuration="1.766624455s" podCreationTimestamp="2025-11-25 10:07:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:07:09.762867081 +0000 UTC m=+1126.446823508" watchObservedRunningTime="2025-11-25 10:07:09.766624455 +0000 UTC m=+1126.450580872" Nov 25 10:07:09 crc kubenswrapper[4926]: I1125 10:07:09.781262 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-dded-account-create-scxpg"] Nov 25 10:07:10 crc kubenswrapper[4926]: E1125 10:07:10.355569 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81121ebf_1d21_43de_8787_efc9b79578e2.slice/crio-2b44fa0a33b70b579973a762a3b624426c4acdea5da29fc400116e50d1330c67.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.744609 4926 generic.go:334] "Generic (PLEG): container finished" podID="81121ebf-1d21-43de-8787-efc9b79578e2" containerID="2b44fa0a33b70b579973a762a3b624426c4acdea5da29fc400116e50d1330c67" exitCode=0 Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.744709 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f822-account-create-hghgh" event={"ID":"81121ebf-1d21-43de-8787-efc9b79578e2","Type":"ContainerDied","Data":"2b44fa0a33b70b579973a762a3b624426c4acdea5da29fc400116e50d1330c67"} Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.744889 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f822-account-create-hghgh" event={"ID":"81121ebf-1d21-43de-8787-efc9b79578e2","Type":"ContainerStarted","Data":"29590fd09d2ae208f2fb05be35d7c77567ea59a55c0e98a88884c8494b6d25cb"} Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.747122 4926 generic.go:334] "Generic (PLEG): container finished" podID="c7d24462-3c87-4454-bf92-e97345e32e12" containerID="e01be87fd1511b3fab3ca6d93a3ddcc5e8dee7080df81c037dfd317c9d2e6750" exitCode=0 Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.747203 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0b02-account-create-jpmcd" event={"ID":"c7d24462-3c87-4454-bf92-e97345e32e12","Type":"ContainerDied","Data":"e01be87fd1511b3fab3ca6d93a3ddcc5e8dee7080df81c037dfd317c9d2e6750"} Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.749503 4926 generic.go:334] "Generic (PLEG): container finished" podID="c58cf150-5525-4d1c-9376-a3bd78100269" containerID="75e798cfda64009187e2d64cf60d99cae626e1b9a49900c8cb226074133c8fde" exitCode=0 Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.749737 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w4qxc" event={"ID":"c58cf150-5525-4d1c-9376-a3bd78100269","Type":"ContainerDied","Data":"75e798cfda64009187e2d64cf60d99cae626e1b9a49900c8cb226074133c8fde"} Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.757354 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e7b372d-b0e0-4423-9734-00d223b0e99e" containerID="5c92d628d09875a45b90e71d2da83c4d2f7da2dfed7704d011f9786db175b20c" exitCode=0 Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.757600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dded-account-create-scxpg" event={"ID":"2e7b372d-b0e0-4423-9734-00d223b0e99e","Type":"ContainerDied","Data":"5c92d628d09875a45b90e71d2da83c4d2f7da2dfed7704d011f9786db175b20c"} Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.757639 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dded-account-create-scxpg" event={"ID":"2e7b372d-b0e0-4423-9734-00d223b0e99e","Type":"ContainerStarted","Data":"9a94230e439d2d42076f99599d216ca865063b8f4ecb28be8aeb981c4ac18786"} Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.760508 4926 generic.go:334] "Generic (PLEG): container finished" podID="8e32ccd7-0d61-4328-a22b-4637bd3f16f4" containerID="893aa06815f5c11ee726bd49637d5a109d184e83b517cc17a67ea83e5e444eb5" exitCode=0 Nov 25 10:07:10 crc kubenswrapper[4926]: I1125 10:07:10.760591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jqn4l" event={"ID":"8e32ccd7-0d61-4328-a22b-4637bd3f16f4","Type":"ContainerDied","Data":"893aa06815f5c11ee726bd49637d5a109d184e83b517cc17a67ea83e5e444eb5"} Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.058454 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.138000 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.255564 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-operator-scripts\") pod \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.255620 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rsd5\" (UniqueName: \"kubernetes.io/projected/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-kube-api-access-5rsd5\") pod \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\" (UID: \"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc\") " Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.256405 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" (UID: "1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.256761 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.260780 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-kube-api-access-5rsd5" (OuterVolumeSpecName: "kube-api-access-5rsd5") pod "1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" (UID: "1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc"). InnerVolumeSpecName "kube-api-access-5rsd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.358175 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rsd5\" (UniqueName: \"kubernetes.io/projected/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc-kube-api-access-5rsd5\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.780957 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pr557" event={"ID":"1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc","Type":"ContainerDied","Data":"6cde52f52c088df59fd08dbb049d7106917bfde73d816690277e0ad6f85762a6"} Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.782429 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cde52f52c088df59fd08dbb049d7106917bfde73d816690277e0ad6f85762a6" Nov 25 10:07:11 crc kubenswrapper[4926]: I1125 10:07:11.781065 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pr557" Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.025237 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.025599 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.025886 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.026666 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bdee8ff08446a7fcbf9166a91fe5650194045a7c04583b6368b152639d328c44"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.026734 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://bdee8ff08446a7fcbf9166a91fe5650194045a7c04583b6368b152639d328c44" gracePeriod=600 Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.812873 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="bdee8ff08446a7fcbf9166a91fe5650194045a7c04583b6368b152639d328c44" exitCode=0 Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.812928 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"bdee8ff08446a7fcbf9166a91fe5650194045a7c04583b6368b152639d328c44"} Nov 25 10:07:14 crc kubenswrapper[4926]: I1125 10:07:14.812985 4926 scope.go:117] "RemoveContainer" containerID="9bc7b6ba06aa7d0baa2979a51d5458380a1834ddc41f7963299878af0ce453ec" Nov 25 10:07:16 crc kubenswrapper[4926]: I1125 10:07:16.294712 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.705518 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.762874 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.777404 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.790700 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7b372d-b0e0-4423-9734-00d223b0e99e-operator-scripts\") pod \"2e7b372d-b0e0-4423-9734-00d223b0e99e\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.790779 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sckk\" (UniqueName: \"kubernetes.io/projected/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-kube-api-access-7sckk\") pod \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.790811 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-operator-scripts\") pod \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\" (UID: \"8e32ccd7-0d61-4328-a22b-4637bd3f16f4\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.790858 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkdhb\" (UniqueName: \"kubernetes.io/projected/2e7b372d-b0e0-4423-9734-00d223b0e99e-kube-api-access-lkdhb\") pod \"2e7b372d-b0e0-4423-9734-00d223b0e99e\" (UID: \"2e7b372d-b0e0-4423-9734-00d223b0e99e\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.791781 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2e7b372d-b0e0-4423-9734-00d223b0e99e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2e7b372d-b0e0-4423-9734-00d223b0e99e" (UID: "2e7b372d-b0e0-4423-9734-00d223b0e99e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.792294 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e32ccd7-0d61-4328-a22b-4637bd3f16f4" (UID: "8e32ccd7-0d61-4328-a22b-4637bd3f16f4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.793567 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.800935 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e7b372d-b0e0-4423-9734-00d223b0e99e-kube-api-access-lkdhb" (OuterVolumeSpecName: "kube-api-access-lkdhb") pod "2e7b372d-b0e0-4423-9734-00d223b0e99e" (UID: "2e7b372d-b0e0-4423-9734-00d223b0e99e"). InnerVolumeSpecName "kube-api-access-lkdhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.817092 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-kube-api-access-7sckk" (OuterVolumeSpecName: "kube-api-access-7sckk") pod "8e32ccd7-0d61-4328-a22b-4637bd3f16f4" (UID: "8e32ccd7-0d61-4328-a22b-4637bd3f16f4"). InnerVolumeSpecName "kube-api-access-7sckk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.853864 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0b02-account-create-jpmcd" event={"ID":"c7d24462-3c87-4454-bf92-e97345e32e12","Type":"ContainerDied","Data":"99e2b9640d8308643197259e479d8eb156645bea857f5f025242944ad8d5eb94"} Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.853904 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0b02-account-create-jpmcd" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.853910 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99e2b9640d8308643197259e479d8eb156645bea857f5f025242944ad8d5eb94" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.855562 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-w4qxc" event={"ID":"c58cf150-5525-4d1c-9376-a3bd78100269","Type":"ContainerDied","Data":"05712ab1d79bb1d97c2ac750e2ceedf1aa5bbcfe700684b4eef7b254d0ec6b8d"} Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.855612 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="05712ab1d79bb1d97c2ac750e2ceedf1aa5bbcfe700684b4eef7b254d0ec6b8d" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.856717 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-dded-account-create-scxpg" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.856727 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-dded-account-create-scxpg" event={"ID":"2e7b372d-b0e0-4423-9734-00d223b0e99e","Type":"ContainerDied","Data":"9a94230e439d2d42076f99599d216ca865063b8f4ecb28be8aeb981c4ac18786"} Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.856760 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a94230e439d2d42076f99599d216ca865063b8f4ecb28be8aeb981c4ac18786" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.858242 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-jqn4l" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.858231 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-jqn4l" event={"ID":"8e32ccd7-0d61-4328-a22b-4637bd3f16f4","Type":"ContainerDied","Data":"f6b1bd10fe9d54d9044a12274362de00201a0cd2b2de963a26682b3c52b108f6"} Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.858349 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6b1bd10fe9d54d9044a12274362de00201a0cd2b2de963a26682b3c52b108f6" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.860386 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f822-account-create-hghgh" event={"ID":"81121ebf-1d21-43de-8787-efc9b79578e2","Type":"ContainerDied","Data":"29590fd09d2ae208f2fb05be35d7c77567ea59a55c0e98a88884c8494b6d25cb"} Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.860414 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29590fd09d2ae208f2fb05be35d7c77567ea59a55c0e98a88884c8494b6d25cb" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.870453 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.888082 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.893041 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pjzv\" (UniqueName: \"kubernetes.io/projected/c58cf150-5525-4d1c-9376-a3bd78100269-kube-api-access-4pjzv\") pod \"c58cf150-5525-4d1c-9376-a3bd78100269\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.893109 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c58cf150-5525-4d1c-9376-a3bd78100269-operator-scripts\") pod \"c58cf150-5525-4d1c-9376-a3bd78100269\" (UID: \"c58cf150-5525-4d1c-9376-a3bd78100269\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.893173 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rtc7\" (UniqueName: \"kubernetes.io/projected/c7d24462-3c87-4454-bf92-e97345e32e12-kube-api-access-5rtc7\") pod \"c7d24462-3c87-4454-bf92-e97345e32e12\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.893208 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d24462-3c87-4454-bf92-e97345e32e12-operator-scripts\") pod \"c7d24462-3c87-4454-bf92-e97345e32e12\" (UID: \"c7d24462-3c87-4454-bf92-e97345e32e12\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.893869 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2e7b372d-b0e0-4423-9734-00d223b0e99e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.893929 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sckk\" (UniqueName: \"kubernetes.io/projected/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-kube-api-access-7sckk\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.894022 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e32ccd7-0d61-4328-a22b-4637bd3f16f4-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.894039 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkdhb\" (UniqueName: \"kubernetes.io/projected/2e7b372d-b0e0-4423-9734-00d223b0e99e-kube-api-access-lkdhb\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.894529 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c58cf150-5525-4d1c-9376-a3bd78100269-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c58cf150-5525-4d1c-9376-a3bd78100269" (UID: "c58cf150-5525-4d1c-9376-a3bd78100269"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.894725 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7d24462-3c87-4454-bf92-e97345e32e12-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c7d24462-3c87-4454-bf92-e97345e32e12" (UID: "c7d24462-3c87-4454-bf92-e97345e32e12"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.901264 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58cf150-5525-4d1c-9376-a3bd78100269-kube-api-access-4pjzv" (OuterVolumeSpecName: "kube-api-access-4pjzv") pod "c58cf150-5525-4d1c-9376-a3bd78100269" (UID: "c58cf150-5525-4d1c-9376-a3bd78100269"). InnerVolumeSpecName "kube-api-access-4pjzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.915621 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7d24462-3c87-4454-bf92-e97345e32e12-kube-api-access-5rtc7" (OuterVolumeSpecName: "kube-api-access-5rtc7") pod "c7d24462-3c87-4454-bf92-e97345e32e12" (UID: "c7d24462-3c87-4454-bf92-e97345e32e12"). InnerVolumeSpecName "kube-api-access-5rtc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.994807 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hdb7g\" (UniqueName: \"kubernetes.io/projected/81121ebf-1d21-43de-8787-efc9b79578e2-kube-api-access-hdb7g\") pod \"81121ebf-1d21-43de-8787-efc9b79578e2\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.995125 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81121ebf-1d21-43de-8787-efc9b79578e2-operator-scripts\") pod \"81121ebf-1d21-43de-8787-efc9b79578e2\" (UID: \"81121ebf-1d21-43de-8787-efc9b79578e2\") " Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.995459 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pjzv\" (UniqueName: \"kubernetes.io/projected/c58cf150-5525-4d1c-9376-a3bd78100269-kube-api-access-4pjzv\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.995482 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c58cf150-5525-4d1c-9376-a3bd78100269-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.995494 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rtc7\" (UniqueName: \"kubernetes.io/projected/c7d24462-3c87-4454-bf92-e97345e32e12-kube-api-access-5rtc7\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.995507 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c7d24462-3c87-4454-bf92-e97345e32e12-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.995542 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81121ebf-1d21-43de-8787-efc9b79578e2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81121ebf-1d21-43de-8787-efc9b79578e2" (UID: "81121ebf-1d21-43de-8787-efc9b79578e2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:07:17 crc kubenswrapper[4926]: I1125 10:07:17.998364 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81121ebf-1d21-43de-8787-efc9b79578e2-kube-api-access-hdb7g" (OuterVolumeSpecName: "kube-api-access-hdb7g") pod "81121ebf-1d21-43de-8787-efc9b79578e2" (UID: "81121ebf-1d21-43de-8787-efc9b79578e2"). InnerVolumeSpecName "kube-api-access-hdb7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.098056 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hdb7g\" (UniqueName: \"kubernetes.io/projected/81121ebf-1d21-43de-8787-efc9b79578e2-kube-api-access-hdb7g\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.098085 4926 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81121ebf-1d21-43de-8787-efc9b79578e2-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.644945 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.645259 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-central-agent" containerID="cri-o://68f633a1421878cb98b424fad8270c42f7e0f2fcf2e3d73230c72d505c3085c5" gracePeriod=30 Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.645340 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="sg-core" containerID="cri-o://38a394d53d549a65130a94247267d96b77401d1e71fbe2911945a62931f9afc5" gracePeriod=30 Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.645368 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-notification-agent" containerID="cri-o://b7c5749cefc782238ff9e36132c6c640b89c0635cfa61e518ca3d56aeddb4bfb" gracePeriod=30 Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.645508 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="proxy-httpd" containerID="cri-o://df00c8c47c20fcb0c08de9a847ce27e3a0f4d65adfc8cbca8609d49a2eec630e" gracePeriod=30 Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.873518 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"02f6bba3-15f5-40ad-9624-deb63a912775","Type":"ContainerStarted","Data":"c83e70e573cf0f3c2c02e1e5d82915a32323d83da4a7d35df2caf3728c2dce81"} Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.875769 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"780e5b5fae7c7161700dd19e5dd8229ab43351e7385e07fc9c5c1c4ee3eb5f66"} Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.878167 4926 generic.go:334] "Generic (PLEG): container finished" podID="d1f2720a-802e-404e-9a13-a81db6466073" containerID="df00c8c47c20fcb0c08de9a847ce27e3a0f4d65adfc8cbca8609d49a2eec630e" exitCode=0 Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.878192 4926 generic.go:334] "Generic (PLEG): container finished" podID="d1f2720a-802e-404e-9a13-a81db6466073" containerID="38a394d53d549a65130a94247267d96b77401d1e71fbe2911945a62931f9afc5" exitCode=2 Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.878238 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerDied","Data":"df00c8c47c20fcb0c08de9a847ce27e3a0f4d65adfc8cbca8609d49a2eec630e"} Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.878275 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-w4qxc" Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.878282 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerDied","Data":"38a394d53d549a65130a94247267d96b77401d1e71fbe2911945a62931f9afc5"} Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.878254 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f822-account-create-hghgh" Nov 25 10:07:18 crc kubenswrapper[4926]: I1125 10:07:18.897980 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.401453307 podStartE2EDuration="14.897946859s" podCreationTimestamp="2025-11-25 10:07:04 +0000 UTC" firstStartedPulling="2025-11-25 10:07:05.067184037 +0000 UTC m=+1121.751140474" lastFinishedPulling="2025-11-25 10:07:17.563677599 +0000 UTC m=+1134.247634026" observedRunningTime="2025-11-25 10:07:18.887348925 +0000 UTC m=+1135.571305342" watchObservedRunningTime="2025-11-25 10:07:18.897946859 +0000 UTC m=+1135.581903286" Nov 25 10:07:19 crc kubenswrapper[4926]: I1125 10:07:19.889239 4926 generic.go:334] "Generic (PLEG): container finished" podID="d1f2720a-802e-404e-9a13-a81db6466073" containerID="68f633a1421878cb98b424fad8270c42f7e0f2fcf2e3d73230c72d505c3085c5" exitCode=0 Nov 25 10:07:19 crc kubenswrapper[4926]: I1125 10:07:19.889323 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerDied","Data":"68f633a1421878cb98b424fad8270c42f7e0f2fcf2e3d73230c72d505c3085c5"} Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.854252 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.879845 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-58nbk\" (UniqueName: \"kubernetes.io/projected/0405b7fb-14c1-4958-a25a-560bfd8bf859-kube-api-access-58nbk\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.879914 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0405b7fb-14c1-4958-a25a-560bfd8bf859-etc-machine-id\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.879968 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0405b7fb-14c1-4958-a25a-560bfd8bf859-logs\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880020 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880040 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0405b7fb-14c1-4958-a25a-560bfd8bf859-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880075 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-combined-ca-bundle\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880119 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-scripts\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880143 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data-custom\") pod \"0405b7fb-14c1-4958-a25a-560bfd8bf859\" (UID: \"0405b7fb-14c1-4958-a25a-560bfd8bf859\") " Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880456 4926 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0405b7fb-14c1-4958-a25a-560bfd8bf859-etc-machine-id\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.880578 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0405b7fb-14c1-4958-a25a-560bfd8bf859-logs" (OuterVolumeSpecName: "logs") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.888499 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.893417 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-scripts" (OuterVolumeSpecName: "scripts") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.901231 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0405b7fb-14c1-4958-a25a-560bfd8bf859-kube-api-access-58nbk" (OuterVolumeSpecName: "kube-api-access-58nbk") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "kube-api-access-58nbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.917350 4926 generic.go:334] "Generic (PLEG): container finished" podID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerID="6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b" exitCode=137 Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.917688 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.917718 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0405b7fb-14c1-4958-a25a-560bfd8bf859","Type":"ContainerDied","Data":"6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b"} Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.917774 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"0405b7fb-14c1-4958-a25a-560bfd8bf859","Type":"ContainerDied","Data":"56c3c225ea728ccc6748fb71702ebc52180f7b9f474a77191c619ca04ac437a3"} Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.917809 4926 scope.go:117] "RemoveContainer" containerID="6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.930539 4926 generic.go:334] "Generic (PLEG): container finished" podID="d1f2720a-802e-404e-9a13-a81db6466073" containerID="b7c5749cefc782238ff9e36132c6c640b89c0635cfa61e518ca3d56aeddb4bfb" exitCode=0 Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.930577 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerDied","Data":"b7c5749cefc782238ff9e36132c6c640b89c0635cfa61e518ca3d56aeddb4bfb"} Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.937856 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.981477 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.981504 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.981512 4926 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data-custom\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.981520 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-58nbk\" (UniqueName: \"kubernetes.io/projected/0405b7fb-14c1-4958-a25a-560bfd8bf859-kube-api-access-58nbk\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.981529 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0405b7fb-14c1-4958-a25a-560bfd8bf859-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:22 crc kubenswrapper[4926]: I1125 10:07:22.999633 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data" (OuterVolumeSpecName: "config-data") pod "0405b7fb-14c1-4958-a25a-560bfd8bf859" (UID: "0405b7fb-14c1-4958-a25a-560bfd8bf859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.028072 4926 scope.go:117] "RemoveContainer" containerID="347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.054928 4926 scope.go:117] "RemoveContainer" containerID="6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.057205 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b\": container with ID starting with 6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b not found: ID does not exist" containerID="6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.057240 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b"} err="failed to get container status \"6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b\": rpc error: code = NotFound desc = could not find container \"6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b\": container with ID starting with 6bf7dae553b7513bde9f604fa985bb15237dfce0c0e8f2019fa0af375abe037b not found: ID does not exist" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.057264 4926 scope.go:117] "RemoveContainer" containerID="347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.057460 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8\": container with ID starting with 347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8 not found: ID does not exist" containerID="347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.057484 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8"} err="failed to get container status \"347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8\": rpc error: code = NotFound desc = could not find container \"347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8\": container with ID starting with 347fbf0a1bee4337db2f3e8e31f688ea70102ddb6fe36ed2a14a7bc1bd5a89d8 not found: ID does not exist" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.086641 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0405b7fb-14c1-4958-a25a-560bfd8bf859-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.089363 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.187778 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-scripts\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.187861 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-combined-ca-bundle\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.187894 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-sg-core-conf-yaml\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.187926 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-run-httpd\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.187987 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-config-data\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.188038 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tdl7m\" (UniqueName: \"kubernetes.io/projected/d1f2720a-802e-404e-9a13-a81db6466073-kube-api-access-tdl7m\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.188155 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-log-httpd\") pod \"d1f2720a-802e-404e-9a13-a81db6466073\" (UID: \"d1f2720a-802e-404e-9a13-a81db6466073\") " Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.189021 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.193415 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.196337 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-scripts" (OuterVolumeSpecName: "scripts") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.207119 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1f2720a-802e-404e-9a13-a81db6466073-kube-api-access-tdl7m" (OuterVolumeSpecName: "kube-api-access-tdl7m") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "kube-api-access-tdl7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.223650 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.293841 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.293877 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.293889 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.293902 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tdl7m\" (UniqueName: \"kubernetes.io/projected/d1f2720a-802e-404e-9a13-a81db6466073-kube-api-access-tdl7m\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.293912 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d1f2720a-802e-404e-9a13-a81db6466073-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.329680 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.370547 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.373654 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-config-data" (OuterVolumeSpecName: "config-data") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378461 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378759 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7d24462-3c87-4454-bf92-e97345e32e12" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378770 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7d24462-3c87-4454-bf92-e97345e32e12" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378788 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-notification-agent" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378795 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-notification-agent" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378807 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e7b372d-b0e0-4423-9734-00d223b0e99e" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378813 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e7b372d-b0e0-4423-9734-00d223b0e99e" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378826 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378832 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378845 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api-log" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378851 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api-log" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378860 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81121ebf-1d21-43de-8787-efc9b79578e2" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378865 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="81121ebf-1d21-43de-8787-efc9b79578e2" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378879 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e32ccd7-0d61-4328-a22b-4637bd3f16f4" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378884 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e32ccd7-0d61-4328-a22b-4637bd3f16f4" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378895 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-central-agent" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378901 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-central-agent" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378910 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378916 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378927 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="proxy-httpd" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378935 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="proxy-httpd" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378965 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="sg-core" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378971 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="sg-core" Nov 25 10:07:23 crc kubenswrapper[4926]: E1125 10:07:23.378981 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58cf150-5525-4d1c-9376-a3bd78100269" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.378987 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58cf150-5525-4d1c-9376-a3bd78100269" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379150 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api-log" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379164 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="proxy-httpd" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379176 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e32ccd7-0d61-4328-a22b-4637bd3f16f4" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379183 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58cf150-5525-4d1c-9376-a3bd78100269" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379194 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7d24462-3c87-4454-bf92-e97345e32e12" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379203 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-notification-agent" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379213 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="81121ebf-1d21-43de-8787-efc9b79578e2" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379224 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e7b372d-b0e0-4423-9734-00d223b0e99e" containerName="mariadb-account-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379233 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" containerName="cinder-api" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379241 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="sg-core" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379248 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1f2720a-802e-404e-9a13-a81db6466073" containerName="ceilometer-central-agent" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.379258 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" containerName="mariadb-database-create" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.380157 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.385286 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.385451 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.385611 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.388186 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d1f2720a-802e-404e-9a13-a81db6466073" (UID: "d1f2720a-802e-404e-9a13-a81db6466073"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.395157 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.395190 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d1f2720a-802e-404e-9a13-a81db6466073-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.401835 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496675 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87b02209-e235-4ce3-8f41-39fd50fd81c9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496713 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t4ndf\" (UniqueName: \"kubernetes.io/projected/87b02209-e235-4ce3-8f41-39fd50fd81c9-kube-api-access-t4ndf\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496755 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496793 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-config-data\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496814 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b02209-e235-4ce3-8f41-39fd50fd81c9-logs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496834 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-scripts\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496903 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-config-data-custom\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.496921 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.597900 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-config-data-custom\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.597984 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598031 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87b02209-e235-4ce3-8f41-39fd50fd81c9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598056 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t4ndf\" (UniqueName: \"kubernetes.io/projected/87b02209-e235-4ce3-8f41-39fd50fd81c9-kube-api-access-t4ndf\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598108 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598142 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598170 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-config-data\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598200 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b02209-e235-4ce3-8f41-39fd50fd81c9-logs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.598226 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-scripts\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.601072 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/87b02209-e235-4ce3-8f41-39fd50fd81c9-etc-machine-id\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.603737 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-scripts\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.604092 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87b02209-e235-4ce3-8f41-39fd50fd81c9-logs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.604556 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-public-tls-certs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.605141 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.606425 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="435aea2d-06c8-43f7-a905-9b1f9e2da38f" containerName="kube-state-metrics" containerID="cri-o://255f0b47abd114ce9e1d3b22dcecd051f3fe06d6de05d51e5bedcbc583028585" gracePeriod=30 Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.608707 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-config-data-custom\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.616745 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.617681 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-config-data\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.618572 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87b02209-e235-4ce3-8f41-39fd50fd81c9-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.624646 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t4ndf\" (UniqueName: \"kubernetes.io/projected/87b02209-e235-4ce3-8f41-39fd50fd81c9-kube-api-access-t4ndf\") pod \"cinder-api-0\" (UID: \"87b02209-e235-4ce3-8f41-39fd50fd81c9\") " pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.713635 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.943774 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d1f2720a-802e-404e-9a13-a81db6466073","Type":"ContainerDied","Data":"4266b51673f65de180aaa17891dcc0a449ab3e32c8a15da54758cb5033a739b3"} Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.944140 4926 scope.go:117] "RemoveContainer" containerID="df00c8c47c20fcb0c08de9a847ce27e3a0f4d65adfc8cbca8609d49a2eec630e" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.943923 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.949238 4926 generic.go:334] "Generic (PLEG): container finished" podID="435aea2d-06c8-43f7-a905-9b1f9e2da38f" containerID="255f0b47abd114ce9e1d3b22dcecd051f3fe06d6de05d51e5bedcbc583028585" exitCode=2 Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.949285 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435aea2d-06c8-43f7-a905-9b1f9e2da38f","Type":"ContainerDied","Data":"255f0b47abd114ce9e1d3b22dcecd051f3fe06d6de05d51e5bedcbc583028585"} Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.984686 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:23 crc kubenswrapper[4926]: I1125 10:07:23.985067 4926 scope.go:117] "RemoveContainer" containerID="38a394d53d549a65130a94247267d96b77401d1e71fbe2911945a62931f9afc5" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.010053 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.055694 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0405b7fb-14c1-4958-a25a-560bfd8bf859" path="/var/lib/kubelet/pods/0405b7fb-14c1-4958-a25a-560bfd8bf859/volumes" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.058320 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1f2720a-802e-404e-9a13-a81db6466073" path="/var/lib/kubelet/pods/d1f2720a-802e-404e-9a13-a81db6466073/volumes" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.060167 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.062758 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.062912 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.076801 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.077064 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.087018 4926 scope.go:117] "RemoveContainer" containerID="b7c5749cefc782238ff9e36132c6c640b89c0635cfa61e518ca3d56aeddb4bfb" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.102453 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-twczw"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.103794 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115312 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-config-data\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115385 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115414 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-run-httpd\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115441 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115480 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-scripts\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115513 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c4nc\" (UniqueName: \"kubernetes.io/projected/af805f60-ca16-40d6-9d01-4c63e418c220-kube-api-access-5c4nc\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.115602 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-log-httpd\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.119697 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-twczw"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.120325 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.120734 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gjcvm" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.120957 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.169612 4926 scope.go:117] "RemoveContainer" containerID="68f633a1421878cb98b424fad8270c42f7e0f2fcf2e3d73230c72d505c3085c5" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.173147 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216248 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fc9k\" (UniqueName: \"kubernetes.io/projected/435aea2d-06c8-43f7-a905-9b1f9e2da38f-kube-api-access-6fc9k\") pod \"435aea2d-06c8-43f7-a905-9b1f9e2da38f\" (UID: \"435aea2d-06c8-43f7-a905-9b1f9e2da38f\") " Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-run-httpd\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216611 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216650 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-scripts\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216678 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c4nc\" (UniqueName: \"kubernetes.io/projected/af805f60-ca16-40d6-9d01-4c63e418c220-kube-api-access-5c4nc\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216711 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-scripts\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216728 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfdcp\" (UniqueName: \"kubernetes.io/projected/a33ef41d-bc58-413a-946a-41e0a85450e5-kube-api-access-rfdcp\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216778 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-log-httpd\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216823 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216846 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-config-data\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216870 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-config-data\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.216903 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.217501 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-run-httpd\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.217657 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-log-httpd\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.221926 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/435aea2d-06c8-43f7-a905-9b1f9e2da38f-kube-api-access-6fc9k" (OuterVolumeSpecName: "kube-api-access-6fc9k") pod "435aea2d-06c8-43f7-a905-9b1f9e2da38f" (UID: "435aea2d-06c8-43f7-a905-9b1f9e2da38f"). InnerVolumeSpecName "kube-api-access-6fc9k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.222575 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.222621 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-scripts\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.223180 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-config-data\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.227165 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.238668 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c4nc\" (UniqueName: \"kubernetes.io/projected/af805f60-ca16-40d6-9d01-4c63e418c220-kube-api-access-5c4nc\") pod \"ceilometer-0\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.262388 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Nov 25 10:07:24 crc kubenswrapper[4926]: W1125 10:07:24.263653 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87b02209_e235_4ce3_8f41_39fd50fd81c9.slice/crio-ad9c831fd850079d96ff4fb796efbfe6efd82b304f9d54cdb8c925c94ebfd859 WatchSource:0}: Error finding container ad9c831fd850079d96ff4fb796efbfe6efd82b304f9d54cdb8c925c94ebfd859: Status 404 returned error can't find the container with id ad9c831fd850079d96ff4fb796efbfe6efd82b304f9d54cdb8c925c94ebfd859 Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.320349 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-scripts\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.320425 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfdcp\" (UniqueName: \"kubernetes.io/projected/a33ef41d-bc58-413a-946a-41e0a85450e5-kube-api-access-rfdcp\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.321304 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.321380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-config-data\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.321737 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fc9k\" (UniqueName: \"kubernetes.io/projected/435aea2d-06c8-43f7-a905-9b1f9e2da38f-kube-api-access-6fc9k\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.326199 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.328761 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-scripts\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.328767 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-config-data\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.337877 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfdcp\" (UniqueName: \"kubernetes.io/projected/a33ef41d-bc58-413a-946a-41e0a85450e5-kube-api-access-rfdcp\") pod \"nova-cell0-conductor-db-sync-twczw\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.461358 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.470468 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.812499 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.975572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87b02209-e235-4ce3-8f41-39fd50fd81c9","Type":"ContainerStarted","Data":"ad9c831fd850079d96ff4fb796efbfe6efd82b304f9d54cdb8c925c94ebfd859"} Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.979216 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-twczw"] Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.985047 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"435aea2d-06c8-43f7-a905-9b1f9e2da38f","Type":"ContainerDied","Data":"eac837a6422e6254720852dca60ae55f196ae93fd4eee28cb330469c1da7086a"} Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.985145 4926 scope.go:117] "RemoveContainer" containerID="255f0b47abd114ce9e1d3b22dcecd051f3fe06d6de05d51e5bedcbc583028585" Nov 25 10:07:24 crc kubenswrapper[4926]: I1125 10:07:24.985155 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.043072 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.138328 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.151113 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.164556 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:07:25 crc kubenswrapper[4926]: E1125 10:07:25.164919 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="435aea2d-06c8-43f7-a905-9b1f9e2da38f" containerName="kube-state-metrics" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.164954 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="435aea2d-06c8-43f7-a905-9b1f9e2da38f" containerName="kube-state-metrics" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.165134 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="435aea2d-06c8-43f7-a905-9b1f9e2da38f" containerName="kube-state-metrics" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.165686 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.167828 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.167924 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.172982 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.250221 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.250509 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9xkx\" (UniqueName: \"kubernetes.io/projected/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-api-access-h9xkx\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.250633 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.250851 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.351919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9xkx\" (UniqueName: \"kubernetes.io/projected/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-api-access-h9xkx\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.352091 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.352212 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.352240 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.357314 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.357346 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.357559 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f8efeb7-cb75-4a17-bf87-85431d9e395e-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.374412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9xkx\" (UniqueName: \"kubernetes.io/projected/2f8efeb7-cb75-4a17-bf87-85431d9e395e-kube-api-access-h9xkx\") pod \"kube-state-metrics-0\" (UID: \"2f8efeb7-cb75-4a17-bf87-85431d9e395e\") " pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.489721 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.984820 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.995638 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87b02209-e235-4ce3-8f41-39fd50fd81c9","Type":"ContainerStarted","Data":"5309c07a9bc57797185b48a69db3ef623a50fff5ec79eb1fbb0f5fb697641e9d"} Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.995686 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"87b02209-e235-4ce3-8f41-39fd50fd81c9","Type":"ContainerStarted","Data":"faa822fec5aca6ee105c85f57e6f8d3c5899f1be7aada00f05751d735f70e882"} Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.995735 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.998127 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerStarted","Data":"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a"} Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.998161 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerStarted","Data":"f944a674b18cc36bb2f7180ea157a7709c99cf2f3fe3a4a16ba14071ffdd5c63"} Nov 25 10:07:25 crc kubenswrapper[4926]: I1125 10:07:25.999351 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-twczw" event={"ID":"a33ef41d-bc58-413a-946a-41e0a85450e5","Type":"ContainerStarted","Data":"7b85b989bd8f5ffac2d45a6c9be1e5521825fe99fbf9bdbdb5ae8dd7fdf1597c"} Nov 25 10:07:26 crc kubenswrapper[4926]: I1125 10:07:26.024409 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.024393027 podStartE2EDuration="3.024393027s" podCreationTimestamp="2025-11-25 10:07:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:07:26.013498065 +0000 UTC m=+1142.697454492" watchObservedRunningTime="2025-11-25 10:07:26.024393027 +0000 UTC m=+1142.708349454" Nov 25 10:07:26 crc kubenswrapper[4926]: I1125 10:07:26.034578 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="435aea2d-06c8-43f7-a905-9b1f9e2da38f" path="/var/lib/kubelet/pods/435aea2d-06c8-43f7-a905-9b1f9e2da38f/volumes" Nov 25 10:07:27 crc kubenswrapper[4926]: I1125 10:07:27.015353 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerStarted","Data":"6fc501f19762f6105670f5e405746e249f6b955c4433daa84248acf85c9e864b"} Nov 25 10:07:27 crc kubenswrapper[4926]: I1125 10:07:27.015859 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerStarted","Data":"e9a13bd0cc7a27caf870c6997fbaff0368ab9aa52e0935a5e0024fc35c2d2604"} Nov 25 10:07:27 crc kubenswrapper[4926]: I1125 10:07:27.015886 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:07:27 crc kubenswrapper[4926]: I1125 10:07:27.018998 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerStarted","Data":"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746"} Nov 25 10:07:27 crc kubenswrapper[4926]: I1125 10:07:27.036965 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.60537708 podStartE2EDuration="2.036917477s" podCreationTimestamp="2025-11-25 10:07:25 +0000 UTC" firstStartedPulling="2025-11-25 10:07:25.996524852 +0000 UTC m=+1142.680481279" lastFinishedPulling="2025-11-25 10:07:26.428065249 +0000 UTC m=+1143.112021676" observedRunningTime="2025-11-25 10:07:27.034333893 +0000 UTC m=+1143.718290330" watchObservedRunningTime="2025-11-25 10:07:27.036917477 +0000 UTC m=+1143.720873924" Nov 25 10:07:28 crc kubenswrapper[4926]: I1125 10:07:28.037814 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerStarted","Data":"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544"} Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.091157 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-twczw" event={"ID":"a33ef41d-bc58-413a-946a-41e0a85450e5","Type":"ContainerStarted","Data":"9d5ac90b538ee9d752eedfe0425ee54eaff126c84a7dbaae8e2042a050f42521"} Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.094075 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerStarted","Data":"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156"} Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.094288 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.094307 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-notification-agent" containerID="cri-o://d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" gracePeriod=30 Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.094349 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="sg-core" containerID="cri-o://9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" gracePeriod=30 Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.094275 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="proxy-httpd" containerID="cri-o://dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" gracePeriod=30 Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.094269 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-central-agent" containerID="cri-o://1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" gracePeriod=30 Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.115579 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-twczw" podStartSLOduration=1.5172258429999999 podStartE2EDuration="9.115563394s" podCreationTimestamp="2025-11-25 10:07:24 +0000 UTC" firstStartedPulling="2025-11-25 10:07:24.993457078 +0000 UTC m=+1141.677413505" lastFinishedPulling="2025-11-25 10:07:32.591794629 +0000 UTC m=+1149.275751056" observedRunningTime="2025-11-25 10:07:33.107826111 +0000 UTC m=+1149.791782548" watchObservedRunningTime="2025-11-25 10:07:33.115563394 +0000 UTC m=+1149.799519821" Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.142640 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.681353616 podStartE2EDuration="10.142621659s" podCreationTimestamp="2025-11-25 10:07:23 +0000 UTC" firstStartedPulling="2025-11-25 10:07:25.064597962 +0000 UTC m=+1141.748554389" lastFinishedPulling="2025-11-25 10:07:32.525866005 +0000 UTC m=+1149.209822432" observedRunningTime="2025-11-25 10:07:33.135809979 +0000 UTC m=+1149.819766406" watchObservedRunningTime="2025-11-25 10:07:33.142621659 +0000 UTC m=+1149.826578086" Nov 25 10:07:33 crc kubenswrapper[4926]: I1125 10:07:33.857092 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020073 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-scripts\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020124 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-config-data\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020149 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-run-httpd\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020183 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-combined-ca-bundle\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020246 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-log-httpd\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020326 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c4nc\" (UniqueName: \"kubernetes.io/projected/af805f60-ca16-40d6-9d01-4c63e418c220-kube-api-access-5c4nc\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020361 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-sg-core-conf-yaml\") pod \"af805f60-ca16-40d6-9d01-4c63e418c220\" (UID: \"af805f60-ca16-40d6-9d01-4c63e418c220\") " Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020802 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.020921 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.021325 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.021341 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/af805f60-ca16-40d6-9d01-4c63e418c220-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.034681 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-scripts" (OuterVolumeSpecName: "scripts") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.041768 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af805f60-ca16-40d6-9d01-4c63e418c220-kube-api-access-5c4nc" (OuterVolumeSpecName: "kube-api-access-5c4nc") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "kube-api-access-5c4nc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.051702 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.106462 4926 generic.go:334] "Generic (PLEG): container finished" podID="af805f60-ca16-40d6-9d01-4c63e418c220" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" exitCode=0 Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.106503 4926 generic.go:334] "Generic (PLEG): container finished" podID="af805f60-ca16-40d6-9d01-4c63e418c220" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" exitCode=2 Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.106515 4926 generic.go:334] "Generic (PLEG): container finished" podID="af805f60-ca16-40d6-9d01-4c63e418c220" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" exitCode=0 Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.106528 4926 generic.go:334] "Generic (PLEG): container finished" podID="af805f60-ca16-40d6-9d01-4c63e418c220" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" exitCode=0 Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.106612 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.118748 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerDied","Data":"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156"} Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.118795 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerDied","Data":"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544"} Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.118810 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerDied","Data":"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746"} Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.118819 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerDied","Data":"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a"} Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.118828 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"af805f60-ca16-40d6-9d01-4c63e418c220","Type":"ContainerDied","Data":"f944a674b18cc36bb2f7180ea157a7709c99cf2f3fe3a4a16ba14071ffdd5c63"} Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.118844 4926 scope.go:117] "RemoveContainer" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.120922 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.122638 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.122665 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.122675 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c4nc\" (UniqueName: \"kubernetes.io/projected/af805f60-ca16-40d6-9d01-4c63e418c220-kube-api-access-5c4nc\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.122684 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.140051 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-config-data" (OuterVolumeSpecName: "config-data") pod "af805f60-ca16-40d6-9d01-4c63e418c220" (UID: "af805f60-ca16-40d6-9d01-4c63e418c220"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.142333 4926 scope.go:117] "RemoveContainer" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.165230 4926 scope.go:117] "RemoveContainer" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.199689 4926 scope.go:117] "RemoveContainer" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.222386 4926 scope.go:117] "RemoveContainer" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.222734 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": container with ID starting with dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156 not found: ID does not exist" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.224474 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156"} err="failed to get container status \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": rpc error: code = NotFound desc = could not find container \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": container with ID starting with dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.224512 4926 scope.go:117] "RemoveContainer" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.224030 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/af805f60-ca16-40d6-9d01-4c63e418c220-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.225293 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": container with ID starting with 9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544 not found: ID does not exist" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.225339 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544"} err="failed to get container status \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": rpc error: code = NotFound desc = could not find container \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": container with ID starting with 9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.225370 4926 scope.go:117] "RemoveContainer" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.225729 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": container with ID starting with d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746 not found: ID does not exist" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.225774 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746"} err="failed to get container status \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": rpc error: code = NotFound desc = could not find container \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": container with ID starting with d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.225807 4926 scope.go:117] "RemoveContainer" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.226321 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": container with ID starting with 1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a not found: ID does not exist" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.226356 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a"} err="failed to get container status \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": rpc error: code = NotFound desc = could not find container \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": container with ID starting with 1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.226380 4926 scope.go:117] "RemoveContainer" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.226684 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156"} err="failed to get container status \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": rpc error: code = NotFound desc = could not find container \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": container with ID starting with dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.226713 4926 scope.go:117] "RemoveContainer" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227041 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544"} err="failed to get container status \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": rpc error: code = NotFound desc = could not find container \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": container with ID starting with 9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227073 4926 scope.go:117] "RemoveContainer" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227341 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746"} err="failed to get container status \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": rpc error: code = NotFound desc = could not find container \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": container with ID starting with d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227366 4926 scope.go:117] "RemoveContainer" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227620 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a"} err="failed to get container status \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": rpc error: code = NotFound desc = could not find container \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": container with ID starting with 1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227650 4926 scope.go:117] "RemoveContainer" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227859 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156"} err="failed to get container status \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": rpc error: code = NotFound desc = could not find container \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": container with ID starting with dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.227884 4926 scope.go:117] "RemoveContainer" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.228107 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544"} err="failed to get container status \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": rpc error: code = NotFound desc = could not find container \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": container with ID starting with 9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.228132 4926 scope.go:117] "RemoveContainer" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.228351 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746"} err="failed to get container status \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": rpc error: code = NotFound desc = could not find container \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": container with ID starting with d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.228432 4926 scope.go:117] "RemoveContainer" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.228724 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a"} err="failed to get container status \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": rpc error: code = NotFound desc = could not find container \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": container with ID starting with 1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.228747 4926 scope.go:117] "RemoveContainer" containerID="dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229052 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156"} err="failed to get container status \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": rpc error: code = NotFound desc = could not find container \"dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156\": container with ID starting with dc0d76b00574359bc780536a5dbcb9c0e47a92fd41cda84547b93719111d0156 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229090 4926 scope.go:117] "RemoveContainer" containerID="9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229312 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544"} err="failed to get container status \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": rpc error: code = NotFound desc = could not find container \"9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544\": container with ID starting with 9f0ea47e005d5a77e4b3c69d7f77e6a890db68128bd2170120d13c36ce78a544 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229334 4926 scope.go:117] "RemoveContainer" containerID="d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229577 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746"} err="failed to get container status \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": rpc error: code = NotFound desc = could not find container \"d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746\": container with ID starting with d39333787fd60c0451def7aeca544a5c492ccb5b869ae50ebad0a6404321f746 not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229603 4926 scope.go:117] "RemoveContainer" containerID="1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.229799 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a"} err="failed to get container status \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": rpc error: code = NotFound desc = could not find container \"1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a\": container with ID starting with 1863663aac62229e357cb424e07f6e4f7d43b928ddbc873491297325a202ff7a not found: ID does not exist" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.448410 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.454963 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.483850 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.484323 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="sg-core" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484345 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="sg-core" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.484364 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-notification-agent" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484374 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-notification-agent" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.484409 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-central-agent" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484418 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-central-agent" Nov 25 10:07:34 crc kubenswrapper[4926]: E1125 10:07:34.484430 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="proxy-httpd" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484438 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="proxy-httpd" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484641 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="sg-core" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484664 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-notification-agent" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484681 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="proxy-httpd" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.484700 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" containerName="ceilometer-central-agent" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.486558 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.488687 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.489387 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.491035 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.500569 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.630534 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-config-data\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.630674 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-scripts\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.631076 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-log-httpd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.631193 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.631239 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpbfd\" (UniqueName: \"kubernetes.io/projected/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-kube-api-access-qpbfd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.631266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.631288 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-run-httpd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.631366 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.732728 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-log-httpd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.732791 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.732838 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpbfd\" (UniqueName: \"kubernetes.io/projected/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-kube-api-access-qpbfd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.732875 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.732901 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-run-httpd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.732964 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.733018 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-config-data\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.733068 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-scripts\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.733791 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-run-httpd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.735062 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-log-httpd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.738467 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.739823 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.741682 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-config-data\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.742332 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-scripts\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.743200 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.751675 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpbfd\" (UniqueName: \"kubernetes.io/projected/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-kube-api-access-qpbfd\") pod \"ceilometer-0\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " pod="openstack/ceilometer-0" Nov 25 10:07:34 crc kubenswrapper[4926]: I1125 10:07:34.825346 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:07:35 crc kubenswrapper[4926]: I1125 10:07:35.136144 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:07:35 crc kubenswrapper[4926]: W1125 10:07:35.140521 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3bc01e08_29e2_4a36_b06f_8dac6c33de8e.slice/crio-925fc809ae48226d0d5c0c36241634ccfba2646ba2321ba7d9edaf73788d7000 WatchSource:0}: Error finding container 925fc809ae48226d0d5c0c36241634ccfba2646ba2321ba7d9edaf73788d7000: Status 404 returned error can't find the container with id 925fc809ae48226d0d5c0c36241634ccfba2646ba2321ba7d9edaf73788d7000 Nov 25 10:07:35 crc kubenswrapper[4926]: I1125 10:07:35.498839 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:07:35 crc kubenswrapper[4926]: I1125 10:07:35.719208 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Nov 25 10:07:36 crc kubenswrapper[4926]: I1125 10:07:36.023776 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af805f60-ca16-40d6-9d01-4c63e418c220" path="/var/lib/kubelet/pods/af805f60-ca16-40d6-9d01-4c63e418c220/volumes" Nov 25 10:07:36 crc kubenswrapper[4926]: I1125 10:07:36.133604 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerStarted","Data":"05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4"} Nov 25 10:07:36 crc kubenswrapper[4926]: I1125 10:07:36.133652 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerStarted","Data":"925fc809ae48226d0d5c0c36241634ccfba2646ba2321ba7d9edaf73788d7000"} Nov 25 10:07:37 crc kubenswrapper[4926]: I1125 10:07:37.145037 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerStarted","Data":"78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d"} Nov 25 10:07:41 crc kubenswrapper[4926]: I1125 10:07:41.189027 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerStarted","Data":"3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb"} Nov 25 10:07:43 crc kubenswrapper[4926]: I1125 10:07:43.205517 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerStarted","Data":"ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0"} Nov 25 10:07:43 crc kubenswrapper[4926]: I1125 10:07:43.207125 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:07:43 crc kubenswrapper[4926]: I1125 10:07:43.226613 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.979756503 podStartE2EDuration="9.226599352s" podCreationTimestamp="2025-11-25 10:07:34 +0000 UTC" firstStartedPulling="2025-11-25 10:07:35.144244505 +0000 UTC m=+1151.828200942" lastFinishedPulling="2025-11-25 10:07:42.391087364 +0000 UTC m=+1159.075043791" observedRunningTime="2025-11-25 10:07:43.223866563 +0000 UTC m=+1159.907823020" watchObservedRunningTime="2025-11-25 10:07:43.226599352 +0000 UTC m=+1159.910555779" Nov 25 10:07:46 crc kubenswrapper[4926]: I1125 10:07:46.242225 4926 generic.go:334] "Generic (PLEG): container finished" podID="a33ef41d-bc58-413a-946a-41e0a85450e5" containerID="9d5ac90b538ee9d752eedfe0425ee54eaff126c84a7dbaae8e2042a050f42521" exitCode=0 Nov 25 10:07:46 crc kubenswrapper[4926]: I1125 10:07:46.242635 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-twczw" event={"ID":"a33ef41d-bc58-413a-946a-41e0a85450e5","Type":"ContainerDied","Data":"9d5ac90b538ee9d752eedfe0425ee54eaff126c84a7dbaae8e2042a050f42521"} Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.628632 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.772831 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rfdcp\" (UniqueName: \"kubernetes.io/projected/a33ef41d-bc58-413a-946a-41e0a85450e5-kube-api-access-rfdcp\") pod \"a33ef41d-bc58-413a-946a-41e0a85450e5\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.772974 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-config-data\") pod \"a33ef41d-bc58-413a-946a-41e0a85450e5\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.773131 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-combined-ca-bundle\") pod \"a33ef41d-bc58-413a-946a-41e0a85450e5\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.773189 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-scripts\") pod \"a33ef41d-bc58-413a-946a-41e0a85450e5\" (UID: \"a33ef41d-bc58-413a-946a-41e0a85450e5\") " Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.778357 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-scripts" (OuterVolumeSpecName: "scripts") pod "a33ef41d-bc58-413a-946a-41e0a85450e5" (UID: "a33ef41d-bc58-413a-946a-41e0a85450e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.779098 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a33ef41d-bc58-413a-946a-41e0a85450e5-kube-api-access-rfdcp" (OuterVolumeSpecName: "kube-api-access-rfdcp") pod "a33ef41d-bc58-413a-946a-41e0a85450e5" (UID: "a33ef41d-bc58-413a-946a-41e0a85450e5"). InnerVolumeSpecName "kube-api-access-rfdcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.800796 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-config-data" (OuterVolumeSpecName: "config-data") pod "a33ef41d-bc58-413a-946a-41e0a85450e5" (UID: "a33ef41d-bc58-413a-946a-41e0a85450e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.806408 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a33ef41d-bc58-413a-946a-41e0a85450e5" (UID: "a33ef41d-bc58-413a-946a-41e0a85450e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.875249 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.875281 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.875293 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a33ef41d-bc58-413a-946a-41e0a85450e5-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:47 crc kubenswrapper[4926]: I1125 10:07:47.875301 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rfdcp\" (UniqueName: \"kubernetes.io/projected/a33ef41d-bc58-413a-946a-41e0a85450e5-kube-api-access-rfdcp\") on node \"crc\" DevicePath \"\"" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.272361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-twczw" event={"ID":"a33ef41d-bc58-413a-946a-41e0a85450e5","Type":"ContainerDied","Data":"7b85b989bd8f5ffac2d45a6c9be1e5521825fe99fbf9bdbdb5ae8dd7fdf1597c"} Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.272706 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b85b989bd8f5ffac2d45a6c9be1e5521825fe99fbf9bdbdb5ae8dd7fdf1597c" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.272519 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-twczw" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.382582 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:07:48 crc kubenswrapper[4926]: E1125 10:07:48.383279 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a33ef41d-bc58-413a-946a-41e0a85450e5" containerName="nova-cell0-conductor-db-sync" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.383400 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a33ef41d-bc58-413a-946a-41e0a85450e5" containerName="nova-cell0-conductor-db-sync" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.383722 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a33ef41d-bc58-413a-946a-41e0a85450e5" containerName="nova-cell0-conductor-db-sync" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.384552 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.388269 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gjcvm" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.388375 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.397592 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.488345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqbf6\" (UniqueName: \"kubernetes.io/projected/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-kube-api-access-kqbf6\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.488431 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.488481 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.590219 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqbf6\" (UniqueName: \"kubernetes.io/projected/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-kube-api-access-kqbf6\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.590297 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.590339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.597060 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.605245 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.622659 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqbf6\" (UniqueName: \"kubernetes.io/projected/5bbaac50-a726-4ae4-81b9-cb19e7d2d81a-kube-api-access-kqbf6\") pod \"nova-cell0-conductor-0\" (UID: \"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a\") " pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:48 crc kubenswrapper[4926]: I1125 10:07:48.707888 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:49 crc kubenswrapper[4926]: W1125 10:07:49.149404 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5bbaac50_a726_4ae4_81b9_cb19e7d2d81a.slice/crio-d4a7e3514c7e9ec6a5d062d07b01909b8aa2daf3065de7e07063e30aa3e3e549 WatchSource:0}: Error finding container d4a7e3514c7e9ec6a5d062d07b01909b8aa2daf3065de7e07063e30aa3e3e549: Status 404 returned error can't find the container with id d4a7e3514c7e9ec6a5d062d07b01909b8aa2daf3065de7e07063e30aa3e3e549 Nov 25 10:07:49 crc kubenswrapper[4926]: I1125 10:07:49.150983 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Nov 25 10:07:49 crc kubenswrapper[4926]: I1125 10:07:49.280991 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a","Type":"ContainerStarted","Data":"d4a7e3514c7e9ec6a5d062d07b01909b8aa2daf3065de7e07063e30aa3e3e549"} Nov 25 10:07:50 crc kubenswrapper[4926]: I1125 10:07:50.293872 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"5bbaac50-a726-4ae4-81b9-cb19e7d2d81a","Type":"ContainerStarted","Data":"74671872a934d1b42294edb2a633937f7c86042b401c508343dc411538e8c657"} Nov 25 10:07:50 crc kubenswrapper[4926]: I1125 10:07:50.295013 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:50 crc kubenswrapper[4926]: I1125 10:07:50.317815 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.3177981 podStartE2EDuration="2.3177981s" podCreationTimestamp="2025-11-25 10:07:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:07:50.316792985 +0000 UTC m=+1167.000749422" watchObservedRunningTime="2025-11-25 10:07:50.3177981 +0000 UTC m=+1167.001754547" Nov 25 10:07:58 crc kubenswrapper[4926]: I1125 10:07:58.736431 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.201419 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-xqpfc"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.202797 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.205168 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.206323 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.217493 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.217621 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-scripts\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.217694 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-config-data\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.217816 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cc87k\" (UniqueName: \"kubernetes.io/projected/29ce8c5b-26dd-47bc-b6a4-19e99a313859-kube-api-access-cc87k\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.223613 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xqpfc"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.319460 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.319576 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-scripts\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.319663 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-config-data\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.319767 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cc87k\" (UniqueName: \"kubernetes.io/projected/29ce8c5b-26dd-47bc-b6a4-19e99a313859-kube-api-access-cc87k\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.328270 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.328768 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-scripts\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.333603 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-config-data\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.341311 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.344331 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.347550 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cc87k\" (UniqueName: \"kubernetes.io/projected/29ce8c5b-26dd-47bc-b6a4-19e99a313859-kube-api-access-cc87k\") pod \"nova-cell0-cell-mapping-xqpfc\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.348408 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.392579 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.422423 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-config-data\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.422489 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nc8hb\" (UniqueName: \"kubernetes.io/projected/90c44588-c18e-412d-b23a-771e3e5cbdef-kube-api-access-nc8hb\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.422526 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.422571 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c44588-c18e-412d-b23a-771e3e5cbdef-logs\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.449845 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.451077 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.454470 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.454651 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.462548 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.464746 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.468256 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.513078 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.527952 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c44588-c18e-412d-b23a-771e3e5cbdef-logs\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.528041 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-config-data\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.528096 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nc8hb\" (UniqueName: \"kubernetes.io/projected/90c44588-c18e-412d-b23a-771e3e5cbdef-kube-api-access-nc8hb\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.528134 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.530749 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c44588-c18e-412d-b23a-771e3e5cbdef-logs\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.531174 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.532725 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-config-data\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.561646 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.574203 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nc8hb\" (UniqueName: \"kubernetes.io/projected/90c44588-c18e-412d-b23a-771e3e5cbdef-kube-api-access-nc8hb\") pod \"nova-api-0\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.578612 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-ft7kt"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.580037 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.604592 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.605649 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-ft7kt"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.605718 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.611717 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.632764 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.632882 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-config-data\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.632918 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bswd\" (UniqueName: \"kubernetes.io/projected/476df1c6-913c-4608-95ca-5fab1752684b-kube-api-access-5bswd\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.632971 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12fb535-0caa-4271-b60a-49d7983cee49-logs\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.633020 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.633043 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-config-data\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.633059 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79zrf\" (UniqueName: \"kubernetes.io/projected/f12fb535-0caa-4271-b60a-49d7983cee49-kube-api-access-79zrf\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.634886 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.734698 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735379 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-config-data\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735429 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735463 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bswd\" (UniqueName: \"kubernetes.io/projected/476df1c6-913c-4608-95ca-5fab1752684b-kube-api-access-5bswd\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735486 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7c7m2\" (UniqueName: \"kubernetes.io/projected/b86d5f2e-290b-4856-b454-e86800fdd30e-kube-api-access-7c7m2\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735523 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735540 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-config\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735605 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12fb535-0caa-4271-b60a-49d7983cee49-logs\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735642 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735697 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-config-data\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.735725 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79zrf\" (UniqueName: \"kubernetes.io/projected/f12fb535-0caa-4271-b60a-49d7983cee49-kube-api-access-79zrf\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.736128 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12fb535-0caa-4271-b60a-49d7983cee49-logs\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.736139 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.736234 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.736307 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.736466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnncr\" (UniqueName: \"kubernetes.io/projected/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-kube-api-access-wnncr\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.743787 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-config-data\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.744514 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.744625 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-config-data\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.747338 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.754048 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.754296 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bswd\" (UniqueName: \"kubernetes.io/projected/476df1c6-913c-4608-95ca-5fab1752684b-kube-api-access-5bswd\") pod \"nova-scheduler-0\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.755349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79zrf\" (UniqueName: \"kubernetes.io/projected/f12fb535-0caa-4271-b60a-49d7983cee49-kube-api-access-79zrf\") pod \"nova-metadata-0\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.786199 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.805419 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840322 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840394 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnncr\" (UniqueName: \"kubernetes.io/projected/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-kube-api-access-wnncr\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840433 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840462 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840490 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7c7m2\" (UniqueName: \"kubernetes.io/projected/b86d5f2e-290b-4856-b454-e86800fdd30e-kube-api-access-7c7m2\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840515 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840533 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-config\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.840582 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.842831 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-dns-svc\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.843990 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-nb\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.844317 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-config\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.844901 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.846779 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.850386 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-sb\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.863614 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnncr\" (UniqueName: \"kubernetes.io/projected/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-kube-api-access-wnncr\") pod \"nova-cell1-novncproxy-0\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.886067 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7c7m2\" (UniqueName: \"kubernetes.io/projected/b86d5f2e-290b-4856-b454-e86800fdd30e-kube-api-access-7c7m2\") pod \"dnsmasq-dns-8b8cf6657-ft7kt\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.969300 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:07:59 crc kubenswrapper[4926]: I1125 10:07:59.980498 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.074767 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-xqpfc"] Nov 25 10:08:00 crc kubenswrapper[4926]: W1125 10:08:00.100031 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29ce8c5b_26dd_47bc_b6a4_19e99a313859.slice/crio-e990257007c22c595781d074f59760ea907a8933c689759a56e43011e8bea1a0 WatchSource:0}: Error finding container e990257007c22c595781d074f59760ea907a8933c689759a56e43011e8bea1a0: Status 404 returned error can't find the container with id e990257007c22c595781d074f59760ea907a8933c689759a56e43011e8bea1a0 Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.212259 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lb4mw"] Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.213981 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.218199 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.218419 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.223365 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lb4mw"] Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.250868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqqn4\" (UniqueName: \"kubernetes.io/projected/4c43314e-7f71-40cb-ab8d-8ac5168e5622-kube-api-access-pqqn4\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.250907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-scripts\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.250995 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-config-data\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.251081 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.352747 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-config-data\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.353238 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.353759 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqqn4\" (UniqueName: \"kubernetes.io/projected/4c43314e-7f71-40cb-ab8d-8ac5168e5622-kube-api-access-pqqn4\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.353780 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-scripts\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.360764 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.361218 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-scripts\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.363748 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-config-data\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.377662 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqqn4\" (UniqueName: \"kubernetes.io/projected/4c43314e-7f71-40cb-ab8d-8ac5168e5622-kube-api-access-pqqn4\") pod \"nova-cell1-conductor-db-sync-lb4mw\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.403618 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xqpfc" event={"ID":"29ce8c5b-26dd-47bc-b6a4-19e99a313859","Type":"ContainerStarted","Data":"41b013eed5424ea88277aed24dc581155f1ce52b8c30ca28c8d549b346b4f4ef"} Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.403667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xqpfc" event={"ID":"29ce8c5b-26dd-47bc-b6a4-19e99a313859","Type":"ContainerStarted","Data":"e990257007c22c595781d074f59760ea907a8933c689759a56e43011e8bea1a0"} Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.420614 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-xqpfc" podStartSLOduration=1.420591624 podStartE2EDuration="1.420591624s" podCreationTimestamp="2025-11-25 10:07:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:00.417202299 +0000 UTC m=+1177.101158746" watchObservedRunningTime="2025-11-25 10:08:00.420591624 +0000 UTC m=+1177.104548051" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.442289 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.484207 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.579028 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.603159 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:00 crc kubenswrapper[4926]: W1125 10:08:00.616058 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod476df1c6_913c_4608_95ca_5fab1752684b.slice/crio-3d888f2ccb640b34ece2935cea4ca3f8e06703d70d1e6161b53aa5dfe4aea8c6 WatchSource:0}: Error finding container 3d888f2ccb640b34ece2935cea4ca3f8e06703d70d1e6161b53aa5dfe4aea8c6: Status 404 returned error can't find the container with id 3d888f2ccb640b34ece2935cea4ca3f8e06703d70d1e6161b53aa5dfe4aea8c6 Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.676297 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:00 crc kubenswrapper[4926]: W1125 10:08:00.679774 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode16d795b_bf24_4faa_81b5_d7b1d08bb5a4.slice/crio-cb9891100378b52d9115bed6330bed037e84101f41955568fc14874c77fbf3aa WatchSource:0}: Error finding container cb9891100378b52d9115bed6330bed037e84101f41955568fc14874c77fbf3aa: Status 404 returned error can't find the container with id cb9891100378b52d9115bed6330bed037e84101f41955568fc14874c77fbf3aa Nov 25 10:08:00 crc kubenswrapper[4926]: I1125 10:08:00.693116 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-ft7kt"] Nov 25 10:08:00 crc kubenswrapper[4926]: W1125 10:08:00.703699 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb86d5f2e_290b_4856_b454_e86800fdd30e.slice/crio-4b53ed51f623557b6350c820e2b89f8e90ce04b29ffc19444a8b6da2342f519f WatchSource:0}: Error finding container 4b53ed51f623557b6350c820e2b89f8e90ce04b29ffc19444a8b6da2342f519f: Status 404 returned error can't find the container with id 4b53ed51f623557b6350c820e2b89f8e90ce04b29ffc19444a8b6da2342f519f Nov 25 10:08:01 crc kubenswrapper[4926]: W1125 10:08:01.086104 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c43314e_7f71_40cb_ab8d_8ac5168e5622.slice/crio-cd4f6be22992273db2a0a12af1051d3ccad0b3981c2e0e338fe190c1ab178497 WatchSource:0}: Error finding container cd4f6be22992273db2a0a12af1051d3ccad0b3981c2e0e338fe190c1ab178497: Status 404 returned error can't find the container with id cd4f6be22992273db2a0a12af1051d3ccad0b3981c2e0e338fe190c1ab178497 Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.087677 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lb4mw"] Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.427120 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"476df1c6-913c-4608-95ca-5fab1752684b","Type":"ContainerStarted","Data":"3d888f2ccb640b34ece2935cea4ca3f8e06703d70d1e6161b53aa5dfe4aea8c6"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.429796 4926 generic.go:334] "Generic (PLEG): container finished" podID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerID="5394837af21313f52061f9a04937d79fc8f150ca635e195524a8318d2224d722" exitCode=0 Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.429845 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" event={"ID":"b86d5f2e-290b-4856-b454-e86800fdd30e","Type":"ContainerDied","Data":"5394837af21313f52061f9a04937d79fc8f150ca635e195524a8318d2224d722"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.429865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" event={"ID":"b86d5f2e-290b-4856-b454-e86800fdd30e","Type":"ContainerStarted","Data":"4b53ed51f623557b6350c820e2b89f8e90ce04b29ffc19444a8b6da2342f519f"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.445675 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" event={"ID":"4c43314e-7f71-40cb-ab8d-8ac5168e5622","Type":"ContainerStarted","Data":"67010eccfa60f625268c622a8ac0a3aa8e11f411f5da4021ea963988942a6ca7"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.445734 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" event={"ID":"4c43314e-7f71-40cb-ab8d-8ac5168e5622","Type":"ContainerStarted","Data":"cd4f6be22992273db2a0a12af1051d3ccad0b3981c2e0e338fe190c1ab178497"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.452124 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4","Type":"ContainerStarted","Data":"cb9891100378b52d9115bed6330bed037e84101f41955568fc14874c77fbf3aa"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.462478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f12fb535-0caa-4271-b60a-49d7983cee49","Type":"ContainerStarted","Data":"53a89ea7325ce4f74db35e0c29152acbbc2b930124b3dc49faeb88ea38cb6435"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.469035 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90c44588-c18e-412d-b23a-771e3e5cbdef","Type":"ContainerStarted","Data":"504e95fc49b84f8e8a15e65cd12b887eee3084a5c4936f4586c6cf4124cf9647"} Nov 25 10:08:01 crc kubenswrapper[4926]: I1125 10:08:01.524749 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" podStartSLOduration=1.524721847 podStartE2EDuration="1.524721847s" podCreationTimestamp="2025-11-25 10:08:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:01.514587015 +0000 UTC m=+1178.198543432" watchObservedRunningTime="2025-11-25 10:08:01.524721847 +0000 UTC m=+1178.208678274" Nov 25 10:08:02 crc kubenswrapper[4926]: I1125 10:08:02.478469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" event={"ID":"b86d5f2e-290b-4856-b454-e86800fdd30e","Type":"ContainerStarted","Data":"acdc94c2d20d7b8c6619f73c327ef238d106843c43f5bb29fe9ba111c385f3f3"} Nov 25 10:08:02 crc kubenswrapper[4926]: I1125 10:08:02.479078 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:08:02 crc kubenswrapper[4926]: I1125 10:08:02.507530 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" podStartSLOduration=3.507507076 podStartE2EDuration="3.507507076s" podCreationTimestamp="2025-11-25 10:07:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:02.501116517 +0000 UTC m=+1179.185072944" watchObservedRunningTime="2025-11-25 10:08:02.507507076 +0000 UTC m=+1179.191463493" Nov 25 10:08:02 crc kubenswrapper[4926]: I1125 10:08:02.990855 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:03 crc kubenswrapper[4926]: I1125 10:08:03.000434 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.501292 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90c44588-c18e-412d-b23a-771e3e5cbdef","Type":"ContainerStarted","Data":"ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f"} Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.502240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90c44588-c18e-412d-b23a-771e3e5cbdef","Type":"ContainerStarted","Data":"bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c"} Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.507198 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"476df1c6-913c-4608-95ca-5fab1752684b","Type":"ContainerStarted","Data":"71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409"} Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.510127 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4","Type":"ContainerStarted","Data":"eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632"} Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.510283 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632" gracePeriod=30 Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.517419 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f12fb535-0caa-4271-b60a-49d7983cee49","Type":"ContainerStarted","Data":"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42"} Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.541018 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.34186593 podStartE2EDuration="5.540914095s" podCreationTimestamp="2025-11-25 10:07:59 +0000 UTC" firstStartedPulling="2025-11-25 10:08:00.683413576 +0000 UTC m=+1177.367369993" lastFinishedPulling="2025-11-25 10:08:03.882461731 +0000 UTC m=+1180.566418158" observedRunningTime="2025-11-25 10:08:04.53869426 +0000 UTC m=+1181.222650697" watchObservedRunningTime="2025-11-25 10:08:04.540914095 +0000 UTC m=+1181.224870522" Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.547814 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.119780694 podStartE2EDuration="5.547798797s" podCreationTimestamp="2025-11-25 10:07:59 +0000 UTC" firstStartedPulling="2025-11-25 10:08:00.452468569 +0000 UTC m=+1177.136424986" lastFinishedPulling="2025-11-25 10:08:03.880486662 +0000 UTC m=+1180.564443089" observedRunningTime="2025-11-25 10:08:04.52385643 +0000 UTC m=+1181.207812867" watchObservedRunningTime="2025-11-25 10:08:04.547798797 +0000 UTC m=+1181.231755224" Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.570605 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.31180422 podStartE2EDuration="5.570570064s" podCreationTimestamp="2025-11-25 10:07:59 +0000 UTC" firstStartedPulling="2025-11-25 10:08:00.621724398 +0000 UTC m=+1177.305680855" lastFinishedPulling="2025-11-25 10:08:03.880490272 +0000 UTC m=+1180.564446699" observedRunningTime="2025-11-25 10:08:04.557407406 +0000 UTC m=+1181.241363853" watchObservedRunningTime="2025-11-25 10:08:04.570570064 +0000 UTC m=+1181.254526501" Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.787240 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.852856 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:08:04 crc kubenswrapper[4926]: I1125 10:08:04.981001 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:05 crc kubenswrapper[4926]: I1125 10:08:05.526653 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f12fb535-0caa-4271-b60a-49d7983cee49","Type":"ContainerStarted","Data":"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867"} Nov 25 10:08:05 crc kubenswrapper[4926]: I1125 10:08:05.527151 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-log" containerID="cri-o://a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42" gracePeriod=30 Nov 25 10:08:05 crc kubenswrapper[4926]: I1125 10:08:05.527601 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-metadata" containerID="cri-o://8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867" gracePeriod=30 Nov 25 10:08:05 crc kubenswrapper[4926]: I1125 10:08:05.567410 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.158981988 podStartE2EDuration="6.567366922s" podCreationTimestamp="2025-11-25 10:07:59 +0000 UTC" firstStartedPulling="2025-11-25 10:08:00.47218603 +0000 UTC m=+1177.156142457" lastFinishedPulling="2025-11-25 10:08:03.880570964 +0000 UTC m=+1180.564527391" observedRunningTime="2025-11-25 10:08:05.556144692 +0000 UTC m=+1182.240101139" watchObservedRunningTime="2025-11-25 10:08:05.567366922 +0000 UTC m=+1182.251323359" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.117898 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.217712 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79zrf\" (UniqueName: \"kubernetes.io/projected/f12fb535-0caa-4271-b60a-49d7983cee49-kube-api-access-79zrf\") pod \"f12fb535-0caa-4271-b60a-49d7983cee49\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.219828 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-combined-ca-bundle\") pod \"f12fb535-0caa-4271-b60a-49d7983cee49\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.220125 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12fb535-0caa-4271-b60a-49d7983cee49-logs\") pod \"f12fb535-0caa-4271-b60a-49d7983cee49\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.220178 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-config-data\") pod \"f12fb535-0caa-4271-b60a-49d7983cee49\" (UID: \"f12fb535-0caa-4271-b60a-49d7983cee49\") " Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.220896 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f12fb535-0caa-4271-b60a-49d7983cee49-logs" (OuterVolumeSpecName: "logs") pod "f12fb535-0caa-4271-b60a-49d7983cee49" (UID: "f12fb535-0caa-4271-b60a-49d7983cee49"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.221609 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f12fb535-0caa-4271-b60a-49d7983cee49-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.224355 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f12fb535-0caa-4271-b60a-49d7983cee49-kube-api-access-79zrf" (OuterVolumeSpecName: "kube-api-access-79zrf") pod "f12fb535-0caa-4271-b60a-49d7983cee49" (UID: "f12fb535-0caa-4271-b60a-49d7983cee49"). InnerVolumeSpecName "kube-api-access-79zrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.257791 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f12fb535-0caa-4271-b60a-49d7983cee49" (UID: "f12fb535-0caa-4271-b60a-49d7983cee49"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.268128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-config-data" (OuterVolumeSpecName: "config-data") pod "f12fb535-0caa-4271-b60a-49d7983cee49" (UID: "f12fb535-0caa-4271-b60a-49d7983cee49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.323906 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.323985 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79zrf\" (UniqueName: \"kubernetes.io/projected/f12fb535-0caa-4271-b60a-49d7983cee49-kube-api-access-79zrf\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.323999 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f12fb535-0caa-4271-b60a-49d7983cee49-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543062 4926 generic.go:334] "Generic (PLEG): container finished" podID="f12fb535-0caa-4271-b60a-49d7983cee49" containerID="8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867" exitCode=0 Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543387 4926 generic.go:334] "Generic (PLEG): container finished" podID="f12fb535-0caa-4271-b60a-49d7983cee49" containerID="a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42" exitCode=143 Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543150 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543172 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f12fb535-0caa-4271-b60a-49d7983cee49","Type":"ContainerDied","Data":"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867"} Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543551 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f12fb535-0caa-4271-b60a-49d7983cee49","Type":"ContainerDied","Data":"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42"} Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"f12fb535-0caa-4271-b60a-49d7983cee49","Type":"ContainerDied","Data":"53a89ea7325ce4f74db35e0c29152acbbc2b930124b3dc49faeb88ea38cb6435"} Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.543614 4926 scope.go:117] "RemoveContainer" containerID="8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.571146 4926 scope.go:117] "RemoveContainer" containerID="a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.588705 4926 scope.go:117] "RemoveContainer" containerID="8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867" Nov 25 10:08:06 crc kubenswrapper[4926]: E1125 10:08:06.589141 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867\": container with ID starting with 8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867 not found: ID does not exist" containerID="8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.589176 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867"} err="failed to get container status \"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867\": rpc error: code = NotFound desc = could not find container \"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867\": container with ID starting with 8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867 not found: ID does not exist" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.589202 4926 scope.go:117] "RemoveContainer" containerID="a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42" Nov 25 10:08:06 crc kubenswrapper[4926]: E1125 10:08:06.589651 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42\": container with ID starting with a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42 not found: ID does not exist" containerID="a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.589678 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42"} err="failed to get container status \"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42\": rpc error: code = NotFound desc = could not find container \"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42\": container with ID starting with a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42 not found: ID does not exist" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.589697 4926 scope.go:117] "RemoveContainer" containerID="8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.590230 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867"} err="failed to get container status \"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867\": rpc error: code = NotFound desc = could not find container \"8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867\": container with ID starting with 8ea0cca1ec2c7dfca7890af5154ba9a74b6012c15f7ea883f77873d9095d3867 not found: ID does not exist" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.590252 4926 scope.go:117] "RemoveContainer" containerID="a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.590427 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42"} err="failed to get container status \"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42\": rpc error: code = NotFound desc = could not find container \"a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42\": container with ID starting with a7e2d2bc2c5679db52c504f7b736ea593e4b9120eb66ec0897e695aa5cb00b42 not found: ID does not exist" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.602503 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.612394 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.627794 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:06 crc kubenswrapper[4926]: E1125 10:08:06.628321 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-metadata" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.628344 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-metadata" Nov 25 10:08:06 crc kubenswrapper[4926]: E1125 10:08:06.628363 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-log" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.628372 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-log" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.628613 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-log" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.628649 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" containerName="nova-metadata-metadata" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.629760 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.632245 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.634475 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.635709 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.731723 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wkw4\" (UniqueName: \"kubernetes.io/projected/e9941781-86e0-4b50-93c1-c43ea54b9a9b-kube-api-access-7wkw4\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.731773 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.731828 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-config-data\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.731888 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9941781-86e0-4b50-93c1-c43ea54b9a9b-logs\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.731910 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.833881 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wkw4\" (UniqueName: \"kubernetes.io/projected/e9941781-86e0-4b50-93c1-c43ea54b9a9b-kube-api-access-7wkw4\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.833976 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.834075 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-config-data\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.834148 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9941781-86e0-4b50-93c1-c43ea54b9a9b-logs\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.834178 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.835122 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9941781-86e0-4b50-93c1-c43ea54b9a9b-logs\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.843610 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-config-data\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.843682 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.844097 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.859474 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wkw4\" (UniqueName: \"kubernetes.io/projected/e9941781-86e0-4b50-93c1-c43ea54b9a9b-kube-api-access-7wkw4\") pod \"nova-metadata-0\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " pod="openstack/nova-metadata-0" Nov 25 10:08:06 crc kubenswrapper[4926]: I1125 10:08:06.958167 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:07 crc kubenswrapper[4926]: I1125 10:08:07.452442 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:07 crc kubenswrapper[4926]: I1125 10:08:07.556383 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9941781-86e0-4b50-93c1-c43ea54b9a9b","Type":"ContainerStarted","Data":"70cf86d490a99fa28fdb3e64434ea801c973d2992784b8cee7055d09d3d18b27"} Nov 25 10:08:08 crc kubenswrapper[4926]: I1125 10:08:08.024064 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f12fb535-0caa-4271-b60a-49d7983cee49" path="/var/lib/kubelet/pods/f12fb535-0caa-4271-b60a-49d7983cee49/volumes" Nov 25 10:08:08 crc kubenswrapper[4926]: I1125 10:08:08.568834 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9941781-86e0-4b50-93c1-c43ea54b9a9b","Type":"ContainerStarted","Data":"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec"} Nov 25 10:08:08 crc kubenswrapper[4926]: I1125 10:08:08.568905 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9941781-86e0-4b50-93c1-c43ea54b9a9b","Type":"ContainerStarted","Data":"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2"} Nov 25 10:08:08 crc kubenswrapper[4926]: I1125 10:08:08.602550 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.602529142 podStartE2EDuration="2.602529142s" podCreationTimestamp="2025-11-25 10:08:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:08.592614705 +0000 UTC m=+1185.276571132" watchObservedRunningTime="2025-11-25 10:08:08.602529142 +0000 UTC m=+1185.286485579" Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.587536 4926 generic.go:334] "Generic (PLEG): container finished" podID="4c43314e-7f71-40cb-ab8d-8ac5168e5622" containerID="67010eccfa60f625268c622a8ac0a3aa8e11f411f5da4021ea963988942a6ca7" exitCode=0 Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.588054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" event={"ID":"4c43314e-7f71-40cb-ab8d-8ac5168e5622","Type":"ContainerDied","Data":"67010eccfa60f625268c622a8ac0a3aa8e11f411f5da4021ea963988942a6ca7"} Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.597399 4926 generic.go:334] "Generic (PLEG): container finished" podID="29ce8c5b-26dd-47bc-b6a4-19e99a313859" containerID="41b013eed5424ea88277aed24dc581155f1ce52b8c30ca28c8d549b346b4f4ef" exitCode=0 Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.597525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xqpfc" event={"ID":"29ce8c5b-26dd-47bc-b6a4-19e99a313859","Type":"ContainerDied","Data":"41b013eed5424ea88277aed24dc581155f1ce52b8c30ca28c8d549b346b4f4ef"} Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.748437 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.748540 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.787258 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.834611 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:08:09 crc kubenswrapper[4926]: I1125 10:08:09.971197 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.052332 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-pz6nn"] Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.052847 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerName="dnsmasq-dns" containerID="cri-o://57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f" gracePeriod=10 Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.552479 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.612643 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcw27\" (UniqueName: \"kubernetes.io/projected/44037126-caa3-43d2-abf4-937f10d8fe2a-kube-api-access-tcw27\") pod \"44037126-caa3-43d2-abf4-937f10d8fe2a\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.612968 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-sb\") pod \"44037126-caa3-43d2-abf4-937f10d8fe2a\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.612998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-config\") pod \"44037126-caa3-43d2-abf4-937f10d8fe2a\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.613039 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-dns-svc\") pod \"44037126-caa3-43d2-abf4-937f10d8fe2a\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.613110 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-nb\") pod \"44037126-caa3-43d2-abf4-937f10d8fe2a\" (UID: \"44037126-caa3-43d2-abf4-937f10d8fe2a\") " Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.616162 4926 generic.go:334] "Generic (PLEG): container finished" podID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerID="57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f" exitCode=0 Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.616852 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.617135 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" event={"ID":"44037126-caa3-43d2-abf4-937f10d8fe2a","Type":"ContainerDied","Data":"57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f"} Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.617187 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58db5546cc-pz6nn" event={"ID":"44037126-caa3-43d2-abf4-937f10d8fe2a","Type":"ContainerDied","Data":"5c0cdcd0b912d9e543b102c6c2bb1305eac4ebf5660ff67be06aba4e2a901c39"} Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.617206 4926 scope.go:117] "RemoveContainer" containerID="57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.627673 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44037126-caa3-43d2-abf4-937f10d8fe2a-kube-api-access-tcw27" (OuterVolumeSpecName: "kube-api-access-tcw27") pod "44037126-caa3-43d2-abf4-937f10d8fe2a" (UID: "44037126-caa3-43d2-abf4-937f10d8fe2a"). InnerVolumeSpecName "kube-api-access-tcw27". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.665675 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.673025 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "44037126-caa3-43d2-abf4-937f10d8fe2a" (UID: "44037126-caa3-43d2-abf4-937f10d8fe2a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.682146 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "44037126-caa3-43d2-abf4-937f10d8fe2a" (UID: "44037126-caa3-43d2-abf4-937f10d8fe2a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.693877 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "44037126-caa3-43d2-abf4-937f10d8fe2a" (UID: "44037126-caa3-43d2-abf4-937f10d8fe2a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.715751 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcw27\" (UniqueName: \"kubernetes.io/projected/44037126-caa3-43d2-abf4-937f10d8fe2a-kube-api-access-tcw27\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.715785 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.715796 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.715805 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.752816 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-config" (OuterVolumeSpecName: "config") pod "44037126-caa3-43d2-abf4-937f10d8fe2a" (UID: "44037126-caa3-43d2-abf4-937f10d8fe2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.764045 4926 scope.go:117] "RemoveContainer" containerID="1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.783375 4926 scope.go:117] "RemoveContainer" containerID="57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f" Nov 25 10:08:10 crc kubenswrapper[4926]: E1125 10:08:10.786413 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f\": container with ID starting with 57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f not found: ID does not exist" containerID="57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.786459 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f"} err="failed to get container status \"57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f\": rpc error: code = NotFound desc = could not find container \"57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f\": container with ID starting with 57b35e398407fca279d5edffcc6af26370f9d9a031763b19377beec50ff93c6f not found: ID does not exist" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.786510 4926 scope.go:117] "RemoveContainer" containerID="1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490" Nov 25 10:08:10 crc kubenswrapper[4926]: E1125 10:08:10.786952 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490\": container with ID starting with 1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490 not found: ID does not exist" containerID="1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.786991 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490"} err="failed to get container status \"1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490\": rpc error: code = NotFound desc = could not find container \"1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490\": container with ID starting with 1d140bffdeb0b320dd1098e3d1a74d91bfa252a9ae835674264bdc05a4408490 not found: ID does not exist" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.818621 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44037126-caa3-43d2-abf4-937f10d8fe2a-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.833851 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.833887 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.975776 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-pz6nn"] Nov 25 10:08:10 crc kubenswrapper[4926]: I1125 10:08:10.982914 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58db5546cc-pz6nn"] Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.023928 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.096145 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126119 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-combined-ca-bundle\") pod \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126288 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cc87k\" (UniqueName: \"kubernetes.io/projected/29ce8c5b-26dd-47bc-b6a4-19e99a313859-kube-api-access-cc87k\") pod \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126338 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-combined-ca-bundle\") pod \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126395 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-scripts\") pod \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126428 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-scripts\") pod \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126501 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-config-data\") pod \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126581 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqqn4\" (UniqueName: \"kubernetes.io/projected/4c43314e-7f71-40cb-ab8d-8ac5168e5622-kube-api-access-pqqn4\") pod \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\" (UID: \"4c43314e-7f71-40cb-ab8d-8ac5168e5622\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.126604 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-config-data\") pod \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\" (UID: \"29ce8c5b-26dd-47bc-b6a4-19e99a313859\") " Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.133578 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29ce8c5b-26dd-47bc-b6a4-19e99a313859-kube-api-access-cc87k" (OuterVolumeSpecName: "kube-api-access-cc87k") pod "29ce8c5b-26dd-47bc-b6a4-19e99a313859" (UID: "29ce8c5b-26dd-47bc-b6a4-19e99a313859"). InnerVolumeSpecName "kube-api-access-cc87k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.135502 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-scripts" (OuterVolumeSpecName: "scripts") pod "4c43314e-7f71-40cb-ab8d-8ac5168e5622" (UID: "4c43314e-7f71-40cb-ab8d-8ac5168e5622"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.138344 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-scripts" (OuterVolumeSpecName: "scripts") pod "29ce8c5b-26dd-47bc-b6a4-19e99a313859" (UID: "29ce8c5b-26dd-47bc-b6a4-19e99a313859"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.141144 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c43314e-7f71-40cb-ab8d-8ac5168e5622-kube-api-access-pqqn4" (OuterVolumeSpecName: "kube-api-access-pqqn4") pod "4c43314e-7f71-40cb-ab8d-8ac5168e5622" (UID: "4c43314e-7f71-40cb-ab8d-8ac5168e5622"). InnerVolumeSpecName "kube-api-access-pqqn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.158356 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-config-data" (OuterVolumeSpecName: "config-data") pod "4c43314e-7f71-40cb-ab8d-8ac5168e5622" (UID: "4c43314e-7f71-40cb-ab8d-8ac5168e5622"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.164292 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-config-data" (OuterVolumeSpecName: "config-data") pod "29ce8c5b-26dd-47bc-b6a4-19e99a313859" (UID: "29ce8c5b-26dd-47bc-b6a4-19e99a313859"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.168447 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29ce8c5b-26dd-47bc-b6a4-19e99a313859" (UID: "29ce8c5b-26dd-47bc-b6a4-19e99a313859"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.194452 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4c43314e-7f71-40cb-ab8d-8ac5168e5622" (UID: "4c43314e-7f71-40cb-ab8d-8ac5168e5622"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229367 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229401 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229412 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqqn4\" (UniqueName: \"kubernetes.io/projected/4c43314e-7f71-40cb-ab8d-8ac5168e5622-kube-api-access-pqqn4\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229422 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229430 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29ce8c5b-26dd-47bc-b6a4-19e99a313859-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229438 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cc87k\" (UniqueName: \"kubernetes.io/projected/29ce8c5b-26dd-47bc-b6a4-19e99a313859-kube-api-access-cc87k\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229446 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.229455 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c43314e-7f71-40cb-ab8d-8ac5168e5622-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.636029 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" event={"ID":"4c43314e-7f71-40cb-ab8d-8ac5168e5622","Type":"ContainerDied","Data":"cd4f6be22992273db2a0a12af1051d3ccad0b3981c2e0e338fe190c1ab178497"} Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.636078 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd4f6be22992273db2a0a12af1051d3ccad0b3981c2e0e338fe190c1ab178497" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.636174 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-lb4mw" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.651415 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-xqpfc" event={"ID":"29ce8c5b-26dd-47bc-b6a4-19e99a313859","Type":"ContainerDied","Data":"e990257007c22c595781d074f59760ea907a8933c689759a56e43011e8bea1a0"} Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.651488 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e990257007c22c595781d074f59760ea907a8933c689759a56e43011e8bea1a0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.651620 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-xqpfc" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.751008 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:08:11 crc kubenswrapper[4926]: E1125 10:08:11.752140 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c43314e-7f71-40cb-ab8d-8ac5168e5622" containerName="nova-cell1-conductor-db-sync" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752269 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c43314e-7f71-40cb-ab8d-8ac5168e5622" containerName="nova-cell1-conductor-db-sync" Nov 25 10:08:11 crc kubenswrapper[4926]: E1125 10:08:11.752289 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerName="init" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752297 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerName="init" Nov 25 10:08:11 crc kubenswrapper[4926]: E1125 10:08:11.752332 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="29ce8c5b-26dd-47bc-b6a4-19e99a313859" containerName="nova-manage" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752341 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="29ce8c5b-26dd-47bc-b6a4-19e99a313859" containerName="nova-manage" Nov 25 10:08:11 crc kubenswrapper[4926]: E1125 10:08:11.752359 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerName="dnsmasq-dns" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752373 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerName="dnsmasq-dns" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752605 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="29ce8c5b-26dd-47bc-b6a4-19e99a313859" containerName="nova-manage" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752633 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c43314e-7f71-40cb-ab8d-8ac5168e5622" containerName="nova-cell1-conductor-db-sync" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.752649 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" containerName="dnsmasq-dns" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.753578 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.759015 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.764230 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.824652 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.824902 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-log" containerID="cri-o://bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c" gracePeriod=30 Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.825184 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-api" containerID="cri-o://ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f" gracePeriod=30 Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.837147 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.848364 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spsfq\" (UniqueName: \"kubernetes.io/projected/05c96c75-c327-4561-ac40-59800d92b9a6-kube-api-access-spsfq\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.848407 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c96c75-c327-4561-ac40-59800d92b9a6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.848505 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c96c75-c327-4561-ac40-59800d92b9a6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.885131 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.885354 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-log" containerID="cri-o://a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2" gracePeriod=30 Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.885775 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-metadata" containerID="cri-o://4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec" gracePeriod=30 Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.949836 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spsfq\" (UniqueName: \"kubernetes.io/projected/05c96c75-c327-4561-ac40-59800d92b9a6-kube-api-access-spsfq\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.949895 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c96c75-c327-4561-ac40-59800d92b9a6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.950050 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c96c75-c327-4561-ac40-59800d92b9a6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.955581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05c96c75-c327-4561-ac40-59800d92b9a6-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.955629 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05c96c75-c327-4561-ac40-59800d92b9a6-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:11 crc kubenswrapper[4926]: E1125 10:08:11.956499 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c43314e_7f71_40cb_ab8d_8ac5168e5622.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9941781_86e0_4b50_93c1_c43ea54b9a9b.slice/crio-conmon-a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29ce8c5b_26dd_47bc_b6a4_19e99a313859.slice\": RecentStats: unable to find data in memory cache]" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.958667 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.958720 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:08:11 crc kubenswrapper[4926]: I1125 10:08:11.980612 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spsfq\" (UniqueName: \"kubernetes.io/projected/05c96c75-c327-4561-ac40-59800d92b9a6-kube-api-access-spsfq\") pod \"nova-cell1-conductor-0\" (UID: \"05c96c75-c327-4561-ac40-59800d92b9a6\") " pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.024690 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44037126-caa3-43d2-abf4-937f10d8fe2a" path="/var/lib/kubelet/pods/44037126-caa3-43d2-abf4-937f10d8fe2a/volumes" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.084599 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.554288 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.664522 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-nova-metadata-tls-certs\") pod \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.664585 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9941781-86e0-4b50-93c1-c43ea54b9a9b-logs\") pod \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.664623 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wkw4\" (UniqueName: \"kubernetes.io/projected/e9941781-86e0-4b50-93c1-c43ea54b9a9b-kube-api-access-7wkw4\") pod \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.664909 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-config-data\") pod \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.664986 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-combined-ca-bundle\") pod \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\" (UID: \"e9941781-86e0-4b50-93c1-c43ea54b9a9b\") " Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.665514 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e9941781-86e0-4b50-93c1-c43ea54b9a9b-logs" (OuterVolumeSpecName: "logs") pod "e9941781-86e0-4b50-93c1-c43ea54b9a9b" (UID: "e9941781-86e0-4b50-93c1-c43ea54b9a9b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.667152 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.681217 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e9941781-86e0-4b50-93c1-c43ea54b9a9b-kube-api-access-7wkw4" (OuterVolumeSpecName: "kube-api-access-7wkw4") pod "e9941781-86e0-4b50-93c1-c43ea54b9a9b" (UID: "e9941781-86e0-4b50-93c1-c43ea54b9a9b"). InnerVolumeSpecName "kube-api-access-7wkw4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.692147 4926 generic.go:334] "Generic (PLEG): container finished" podID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerID="bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c" exitCode=143 Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.692225 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90c44588-c18e-412d-b23a-771e3e5cbdef","Type":"ContainerDied","Data":"bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c"} Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.693822 4926 generic.go:334] "Generic (PLEG): container finished" podID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerID="4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec" exitCode=0 Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.693845 4926 generic.go:334] "Generic (PLEG): container finished" podID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerID="a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2" exitCode=143 Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.694074 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="476df1c6-913c-4608-95ca-5fab1752684b" containerName="nova-scheduler-scheduler" containerID="cri-o://71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" gracePeriod=30 Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.694114 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.694138 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9941781-86e0-4b50-93c1-c43ea54b9a9b","Type":"ContainerDied","Data":"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec"} Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.694191 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9941781-86e0-4b50-93c1-c43ea54b9a9b","Type":"ContainerDied","Data":"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2"} Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.694208 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e9941781-86e0-4b50-93c1-c43ea54b9a9b","Type":"ContainerDied","Data":"70cf86d490a99fa28fdb3e64434ea801c973d2992784b8cee7055d09d3d18b27"} Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.694229 4926 scope.go:117] "RemoveContainer" containerID="4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.702864 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-config-data" (OuterVolumeSpecName: "config-data") pod "e9941781-86e0-4b50-93c1-c43ea54b9a9b" (UID: "e9941781-86e0-4b50-93c1-c43ea54b9a9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.710049 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e9941781-86e0-4b50-93c1-c43ea54b9a9b" (UID: "e9941781-86e0-4b50-93c1-c43ea54b9a9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.726152 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e9941781-86e0-4b50-93c1-c43ea54b9a9b" (UID: "e9941781-86e0-4b50-93c1-c43ea54b9a9b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.767507 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.767553 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.767567 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9941781-86e0-4b50-93c1-c43ea54b9a9b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.767579 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9941781-86e0-4b50-93c1-c43ea54b9a9b-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.767590 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wkw4\" (UniqueName: \"kubernetes.io/projected/e9941781-86e0-4b50-93c1-c43ea54b9a9b-kube-api-access-7wkw4\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.775395 4926 scope.go:117] "RemoveContainer" containerID="a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.809447 4926 scope.go:117] "RemoveContainer" containerID="4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec" Nov 25 10:08:12 crc kubenswrapper[4926]: E1125 10:08:12.809948 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec\": container with ID starting with 4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec not found: ID does not exist" containerID="4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.809992 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec"} err="failed to get container status \"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec\": rpc error: code = NotFound desc = could not find container \"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec\": container with ID starting with 4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec not found: ID does not exist" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.810020 4926 scope.go:117] "RemoveContainer" containerID="a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2" Nov 25 10:08:12 crc kubenswrapper[4926]: E1125 10:08:12.810403 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2\": container with ID starting with a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2 not found: ID does not exist" containerID="a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.810446 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2"} err="failed to get container status \"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2\": rpc error: code = NotFound desc = could not find container \"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2\": container with ID starting with a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2 not found: ID does not exist" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.810497 4926 scope.go:117] "RemoveContainer" containerID="4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.810774 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec"} err="failed to get container status \"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec\": rpc error: code = NotFound desc = could not find container \"4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec\": container with ID starting with 4b96113f5917030f76cb6fbd736486227e66b7b32e0990d0d43fc722e0e318ec not found: ID does not exist" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.810797 4926 scope.go:117] "RemoveContainer" containerID="a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2" Nov 25 10:08:12 crc kubenswrapper[4926]: I1125 10:08:12.811041 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2"} err="failed to get container status \"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2\": rpc error: code = NotFound desc = could not find container \"a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2\": container with ID starting with a041fe598975cb803929550bda42c9a5ec04107261a483c1da2f52d19d4e11c2 not found: ID does not exist" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.026138 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.048785 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.057397 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:13 crc kubenswrapper[4926]: E1125 10:08:13.057698 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-metadata" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.057714 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-metadata" Nov 25 10:08:13 crc kubenswrapper[4926]: E1125 10:08:13.057728 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-log" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.057735 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-log" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.057913 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-metadata" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.058012 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" containerName="nova-metadata-log" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.058876 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.061521 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.061846 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.082503 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.177495 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1760dcfa-c766-4bf2-b054-223863000927-logs\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.177553 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvd2f\" (UniqueName: \"kubernetes.io/projected/1760dcfa-c766-4bf2-b054-223863000927-kube-api-access-gvd2f\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.177611 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-config-data\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.177750 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.177839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.278860 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1760dcfa-c766-4bf2-b054-223863000927-logs\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.278912 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvd2f\" (UniqueName: \"kubernetes.io/projected/1760dcfa-c766-4bf2-b054-223863000927-kube-api-access-gvd2f\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.278960 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-config-data\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.279017 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.279080 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.280152 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1760dcfa-c766-4bf2-b054-223863000927-logs\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.284189 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.287761 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-config-data\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.296807 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.297112 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvd2f\" (UniqueName: \"kubernetes.io/projected/1760dcfa-c766-4bf2-b054-223863000927-kube-api-access-gvd2f\") pod \"nova-metadata-0\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.385179 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.703746 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"05c96c75-c327-4561-ac40-59800d92b9a6","Type":"ContainerStarted","Data":"cd1afebc7bf2b585c54f8c1711bd95de4e50fb93bae135b177b5c930c65a52ba"} Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.704115 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.704131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"05c96c75-c327-4561-ac40-59800d92b9a6","Type":"ContainerStarted","Data":"4fbdaf154231dab1def202c3144ffb4b0fad0b46a9400ceabd8210e5efba1c85"} Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.722973 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.722954683 podStartE2EDuration="2.722954683s" podCreationTimestamp="2025-11-25 10:08:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:13.716863411 +0000 UTC m=+1190.400819838" watchObservedRunningTime="2025-11-25 10:08:13.722954683 +0000 UTC m=+1190.406911110" Nov 25 10:08:13 crc kubenswrapper[4926]: I1125 10:08:13.830178 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:14 crc kubenswrapper[4926]: I1125 10:08:14.027788 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e9941781-86e0-4b50-93c1-c43ea54b9a9b" path="/var/lib/kubelet/pods/e9941781-86e0-4b50-93c1-c43ea54b9a9b/volumes" Nov 25 10:08:14 crc kubenswrapper[4926]: I1125 10:08:14.716392 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1760dcfa-c766-4bf2-b054-223863000927","Type":"ContainerStarted","Data":"3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711"} Nov 25 10:08:14 crc kubenswrapper[4926]: I1125 10:08:14.716759 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1760dcfa-c766-4bf2-b054-223863000927","Type":"ContainerStarted","Data":"61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9"} Nov 25 10:08:14 crc kubenswrapper[4926]: I1125 10:08:14.716777 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1760dcfa-c766-4bf2-b054-223863000927","Type":"ContainerStarted","Data":"afb2443e09c16ae6bb12851dcb3f18ea7ff51df1e5feb31d079e602785d29595"} Nov 25 10:08:14 crc kubenswrapper[4926]: I1125 10:08:14.739860 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.739837192 podStartE2EDuration="1.739837192s" podCreationTimestamp="2025-11-25 10:08:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:14.737118235 +0000 UTC m=+1191.421074662" watchObservedRunningTime="2025-11-25 10:08:14.739837192 +0000 UTC m=+1191.423793619" Nov 25 10:08:14 crc kubenswrapper[4926]: E1125 10:08:14.789163 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 10:08:14 crc kubenswrapper[4926]: E1125 10:08:14.790285 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 10:08:14 crc kubenswrapper[4926]: E1125 10:08:14.791500 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Nov 25 10:08:14 crc kubenswrapper[4926]: E1125 10:08:14.791533 4926 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="476df1c6-913c-4608-95ca-5fab1752684b" containerName="nova-scheduler-scheduler" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.343643 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.448753 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-combined-ca-bundle\") pod \"476df1c6-913c-4608-95ca-5fab1752684b\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.448830 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-config-data\") pod \"476df1c6-913c-4608-95ca-5fab1752684b\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.448895 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bswd\" (UniqueName: \"kubernetes.io/projected/476df1c6-913c-4608-95ca-5fab1752684b-kube-api-access-5bswd\") pod \"476df1c6-913c-4608-95ca-5fab1752684b\" (UID: \"476df1c6-913c-4608-95ca-5fab1752684b\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.475731 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/476df1c6-913c-4608-95ca-5fab1752684b-kube-api-access-5bswd" (OuterVolumeSpecName: "kube-api-access-5bswd") pod "476df1c6-913c-4608-95ca-5fab1752684b" (UID: "476df1c6-913c-4608-95ca-5fab1752684b"). InnerVolumeSpecName "kube-api-access-5bswd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.486409 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-config-data" (OuterVolumeSpecName: "config-data") pod "476df1c6-913c-4608-95ca-5fab1752684b" (UID: "476df1c6-913c-4608-95ca-5fab1752684b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.489728 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "476df1c6-913c-4608-95ca-5fab1752684b" (UID: "476df1c6-913c-4608-95ca-5fab1752684b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.550517 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bswd\" (UniqueName: \"kubernetes.io/projected/476df1c6-913c-4608-95ca-5fab1752684b-kube-api-access-5bswd\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.551206 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.551367 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.551392 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/476df1c6-913c-4608-95ca-5fab1752684b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.652693 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nc8hb\" (UniqueName: \"kubernetes.io/projected/90c44588-c18e-412d-b23a-771e3e5cbdef-kube-api-access-nc8hb\") pod \"90c44588-c18e-412d-b23a-771e3e5cbdef\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.652793 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-combined-ca-bundle\") pod \"90c44588-c18e-412d-b23a-771e3e5cbdef\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.652825 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-config-data\") pod \"90c44588-c18e-412d-b23a-771e3e5cbdef\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.652957 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c44588-c18e-412d-b23a-771e3e5cbdef-logs\") pod \"90c44588-c18e-412d-b23a-771e3e5cbdef\" (UID: \"90c44588-c18e-412d-b23a-771e3e5cbdef\") " Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.653493 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90c44588-c18e-412d-b23a-771e3e5cbdef-logs" (OuterVolumeSpecName: "logs") pod "90c44588-c18e-412d-b23a-771e3e5cbdef" (UID: "90c44588-c18e-412d-b23a-771e3e5cbdef"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.656468 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90c44588-c18e-412d-b23a-771e3e5cbdef-kube-api-access-nc8hb" (OuterVolumeSpecName: "kube-api-access-nc8hb") pod "90c44588-c18e-412d-b23a-771e3e5cbdef" (UID: "90c44588-c18e-412d-b23a-771e3e5cbdef"). InnerVolumeSpecName "kube-api-access-nc8hb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.674966 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90c44588-c18e-412d-b23a-771e3e5cbdef" (UID: "90c44588-c18e-412d-b23a-771e3e5cbdef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.681882 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-config-data" (OuterVolumeSpecName: "config-data") pod "90c44588-c18e-412d-b23a-771e3e5cbdef" (UID: "90c44588-c18e-412d-b23a-771e3e5cbdef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.741374 4926 generic.go:334] "Generic (PLEG): container finished" podID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerID="ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f" exitCode=0 Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.741441 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.741445 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90c44588-c18e-412d-b23a-771e3e5cbdef","Type":"ContainerDied","Data":"ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f"} Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.741561 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90c44588-c18e-412d-b23a-771e3e5cbdef","Type":"ContainerDied","Data":"504e95fc49b84f8e8a15e65cd12b887eee3084a5c4936f4586c6cf4124cf9647"} Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.741582 4926 scope.go:117] "RemoveContainer" containerID="ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.742870 4926 generic.go:334] "Generic (PLEG): container finished" podID="476df1c6-913c-4608-95ca-5fab1752684b" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" exitCode=0 Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.742900 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"476df1c6-913c-4608-95ca-5fab1752684b","Type":"ContainerDied","Data":"71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409"} Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.742911 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.742919 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"476df1c6-913c-4608-95ca-5fab1752684b","Type":"ContainerDied","Data":"3d888f2ccb640b34ece2935cea4ca3f8e06703d70d1e6161b53aa5dfe4aea8c6"} Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.754912 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90c44588-c18e-412d-b23a-771e3e5cbdef-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.754950 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nc8hb\" (UniqueName: \"kubernetes.io/projected/90c44588-c18e-412d-b23a-771e3e5cbdef-kube-api-access-nc8hb\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.754960 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.754969 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90c44588-c18e-412d-b23a-771e3e5cbdef-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.766625 4926 scope.go:117] "RemoveContainer" containerID="bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.786686 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.795656 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.798865 4926 scope.go:117] "RemoveContainer" containerID="ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f" Nov 25 10:08:16 crc kubenswrapper[4926]: E1125 10:08:16.799389 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f\": container with ID starting with ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f not found: ID does not exist" containerID="ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.799442 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f"} err="failed to get container status \"ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f\": rpc error: code = NotFound desc = could not find container \"ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f\": container with ID starting with ffaf2647ae47124efd0edfc0ca20e3fe119b9321bf6314b73d26355cc5dd479f not found: ID does not exist" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.799493 4926 scope.go:117] "RemoveContainer" containerID="bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c" Nov 25 10:08:16 crc kubenswrapper[4926]: E1125 10:08:16.799785 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c\": container with ID starting with bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c not found: ID does not exist" containerID="bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.799825 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c"} err="failed to get container status \"bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c\": rpc error: code = NotFound desc = could not find container \"bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c\": container with ID starting with bf87515d079acab3742dfcba0f8b2adcd3b26c9fe85ee99d87497ecea30e8e0c not found: ID does not exist" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.799845 4926 scope.go:117] "RemoveContainer" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.807054 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.861864 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.874765 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: E1125 10:08:16.875281 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-api" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.875300 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-api" Nov 25 10:08:16 crc kubenswrapper[4926]: E1125 10:08:16.875314 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-log" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.875320 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-log" Nov 25 10:08:16 crc kubenswrapper[4926]: E1125 10:08:16.875368 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="476df1c6-913c-4608-95ca-5fab1752684b" containerName="nova-scheduler-scheduler" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.875376 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="476df1c6-913c-4608-95ca-5fab1752684b" containerName="nova-scheduler-scheduler" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.875613 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-log" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.875641 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" containerName="nova-api-api" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.875673 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="476df1c6-913c-4608-95ca-5fab1752684b" containerName="nova-scheduler-scheduler" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.877683 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.879657 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.883655 4926 scope.go:117] "RemoveContainer" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" Nov 25 10:08:16 crc kubenswrapper[4926]: E1125 10:08:16.884106 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409\": container with ID starting with 71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409 not found: ID does not exist" containerID="71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.884142 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409"} err="failed to get container status \"71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409\": rpc error: code = NotFound desc = could not find container \"71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409\": container with ID starting with 71435e792a82a8cbcd0ecc035b25d101053d94696642066ab3b684f8ab5e5409 not found: ID does not exist" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.888404 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.901131 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.902552 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.904578 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.910158 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960404 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2822671e-433a-4983-81cf-c315f9a87184-logs\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960453 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztmz7\" (UniqueName: \"kubernetes.io/projected/2822671e-433a-4983-81cf-c315f9a87184-kube-api-access-ztmz7\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960504 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kb5s\" (UniqueName: \"kubernetes.io/projected/406430ac-1468-495b-a177-52e52f5a0405-kube-api-access-4kb5s\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960561 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-config-data\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960581 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960613 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:16 crc kubenswrapper[4926]: I1125 10:08:16.960666 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-config-data\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.062593 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.062682 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-config-data\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.062797 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2822671e-433a-4983-81cf-c315f9a87184-logs\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.062829 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztmz7\" (UniqueName: \"kubernetes.io/projected/2822671e-433a-4983-81cf-c315f9a87184-kube-api-access-ztmz7\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.062885 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kb5s\" (UniqueName: \"kubernetes.io/projected/406430ac-1468-495b-a177-52e52f5a0405-kube-api-access-4kb5s\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.063017 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-config-data\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.063057 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.063401 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2822671e-433a-4983-81cf-c315f9a87184-logs\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.067297 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.067365 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-config-data\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.067301 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.078315 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-config-data\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.085014 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztmz7\" (UniqueName: \"kubernetes.io/projected/2822671e-433a-4983-81cf-c315f9a87184-kube-api-access-ztmz7\") pod \"nova-api-0\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.127269 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kb5s\" (UniqueName: \"kubernetes.io/projected/406430ac-1468-495b-a177-52e52f5a0405-kube-api-access-4kb5s\") pod \"nova-scheduler-0\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.156647 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.195713 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.224480 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.645711 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:17 crc kubenswrapper[4926]: W1125 10:08:17.652504 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod406430ac_1468_495b_a177_52e52f5a0405.slice/crio-c4e38acc019ca436bf810ddea2c5901d6e90f55cfc78286b774841f43eebc03d WatchSource:0}: Error finding container c4e38acc019ca436bf810ddea2c5901d6e90f55cfc78286b774841f43eebc03d: Status 404 returned error can't find the container with id c4e38acc019ca436bf810ddea2c5901d6e90f55cfc78286b774841f43eebc03d Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.738357 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:17 crc kubenswrapper[4926]: W1125 10:08:17.752124 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2822671e_433a_4983_81cf_c315f9a87184.slice/crio-c41baa2764ab8f086a7a2d3ec812e909f856ebe7590af68182598392f283f379 WatchSource:0}: Error finding container c41baa2764ab8f086a7a2d3ec812e909f856ebe7590af68182598392f283f379: Status 404 returned error can't find the container with id c41baa2764ab8f086a7a2d3ec812e909f856ebe7590af68182598392f283f379 Nov 25 10:08:17 crc kubenswrapper[4926]: I1125 10:08:17.755982 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406430ac-1468-495b-a177-52e52f5a0405","Type":"ContainerStarted","Data":"c4e38acc019ca436bf810ddea2c5901d6e90f55cfc78286b774841f43eebc03d"} Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.024464 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="476df1c6-913c-4608-95ca-5fab1752684b" path="/var/lib/kubelet/pods/476df1c6-913c-4608-95ca-5fab1752684b/volumes" Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.025192 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90c44588-c18e-412d-b23a-771e3e5cbdef" path="/var/lib/kubelet/pods/90c44588-c18e-412d-b23a-771e3e5cbdef/volumes" Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.385297 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.385360 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.769655 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2822671e-433a-4983-81cf-c315f9a87184","Type":"ContainerStarted","Data":"c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322"} Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.769746 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2822671e-433a-4983-81cf-c315f9a87184","Type":"ContainerStarted","Data":"5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d"} Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.770723 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2822671e-433a-4983-81cf-c315f9a87184","Type":"ContainerStarted","Data":"c41baa2764ab8f086a7a2d3ec812e909f856ebe7590af68182598392f283f379"} Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.771020 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406430ac-1468-495b-a177-52e52f5a0405","Type":"ContainerStarted","Data":"f9ad0b36a3f22b70a06652dc863be694e8b4662461d428a18e1ed6a19f3c6ed8"} Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.795836 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.795812899 podStartE2EDuration="2.795812899s" podCreationTimestamp="2025-11-25 10:08:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:18.783488402 +0000 UTC m=+1195.467444839" watchObservedRunningTime="2025-11-25 10:08:18.795812899 +0000 UTC m=+1195.479769336" Nov 25 10:08:18 crc kubenswrapper[4926]: I1125 10:08:18.808673 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.808653859 podStartE2EDuration="2.808653859s" podCreationTimestamp="2025-11-25 10:08:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:18.803213214 +0000 UTC m=+1195.487169661" watchObservedRunningTime="2025-11-25 10:08:18.808653859 +0000 UTC m=+1195.492610286" Nov 25 10:08:22 crc kubenswrapper[4926]: I1125 10:08:22.196442 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:08:23 crc kubenswrapper[4926]: I1125 10:08:23.385819 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:08:23 crc kubenswrapper[4926]: I1125 10:08:23.386217 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:08:24 crc kubenswrapper[4926]: I1125 10:08:24.393425 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:08:24 crc kubenswrapper[4926]: I1125 10:08:24.400636 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:08:27 crc kubenswrapper[4926]: I1125 10:08:27.196972 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:08:27 crc kubenswrapper[4926]: I1125 10:08:27.225643 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:08:27 crc kubenswrapper[4926]: I1125 10:08:27.225700 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:08:27 crc kubenswrapper[4926]: I1125 10:08:27.229385 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:08:27 crc kubenswrapper[4926]: I1125 10:08:27.900820 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:08:28 crc kubenswrapper[4926]: I1125 10:08:28.308220 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:08:28 crc kubenswrapper[4926]: I1125 10:08:28.308211 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.178:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:08:33 crc kubenswrapper[4926]: I1125 10:08:33.390571 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:08:33 crc kubenswrapper[4926]: I1125 10:08:33.393462 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:08:33 crc kubenswrapper[4926]: I1125 10:08:33.397973 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:08:33 crc kubenswrapper[4926]: I1125 10:08:33.924531 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.908128 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.936440 4926 generic.go:334] "Generic (PLEG): container finished" podID="e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" containerID="eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632" exitCode=137 Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.936755 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4","Type":"ContainerDied","Data":"eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632"} Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.936808 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.936845 4926 scope.go:117] "RemoveContainer" containerID="eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632" Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.936829 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4","Type":"ContainerDied","Data":"cb9891100378b52d9115bed6330bed037e84101f41955568fc14874c77fbf3aa"} Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.970349 4926 scope.go:117] "RemoveContainer" containerID="eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632" Nov 25 10:08:34 crc kubenswrapper[4926]: E1125 10:08:34.970844 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632\": container with ID starting with eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632 not found: ID does not exist" containerID="eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632" Nov 25 10:08:34 crc kubenswrapper[4926]: I1125 10:08:34.970878 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632"} err="failed to get container status \"eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632\": rpc error: code = NotFound desc = could not find container \"eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632\": container with ID starting with eb34bc3bc3a68c785e66c21142af24a2913a8c5de2687d042905e695a3300632 not found: ID does not exist" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.109176 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-combined-ca-bundle\") pod \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.109324 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wnncr\" (UniqueName: \"kubernetes.io/projected/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-kube-api-access-wnncr\") pod \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.109363 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-config-data\") pod \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\" (UID: \"e16d795b-bf24-4faa-81b5-d7b1d08bb5a4\") " Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.119091 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-kube-api-access-wnncr" (OuterVolumeSpecName: "kube-api-access-wnncr") pod "e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" (UID: "e16d795b-bf24-4faa-81b5-d7b1d08bb5a4"). InnerVolumeSpecName "kube-api-access-wnncr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.156643 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" (UID: "e16d795b-bf24-4faa-81b5-d7b1d08bb5a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.164003 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-config-data" (OuterVolumeSpecName: "config-data") pod "e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" (UID: "e16d795b-bf24-4faa-81b5-d7b1d08bb5a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.212324 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.212374 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wnncr\" (UniqueName: \"kubernetes.io/projected/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-kube-api-access-wnncr\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.212396 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.311584 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.320557 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.332145 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:35 crc kubenswrapper[4926]: E1125 10:08:35.332655 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.332680 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.333132 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" containerName="nova-cell1-novncproxy-novncproxy" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.334985 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.338591 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.338809 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.340786 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.341203 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.517049 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v7t2\" (UniqueName: \"kubernetes.io/projected/ff3b4281-7815-4363-91b6-fabe18bfdb70-kube-api-access-6v7t2\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.518086 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.518350 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.518585 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.518731 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.621046 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.621099 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.621141 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6v7t2\" (UniqueName: \"kubernetes.io/projected/ff3b4281-7815-4363-91b6-fabe18bfdb70-kube-api-access-6v7t2\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.621228 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.621291 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.625493 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.625684 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.627789 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.628778 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/ff3b4281-7815-4363-91b6-fabe18bfdb70-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.647700 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v7t2\" (UniqueName: \"kubernetes.io/projected/ff3b4281-7815-4363-91b6-fabe18bfdb70-kube-api-access-6v7t2\") pod \"nova-cell1-novncproxy-0\" (UID: \"ff3b4281-7815-4363-91b6-fabe18bfdb70\") " pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:35 crc kubenswrapper[4926]: I1125 10:08:35.659442 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:36 crc kubenswrapper[4926]: I1125 10:08:36.021723 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e16d795b-bf24-4faa-81b5-d7b1d08bb5a4" path="/var/lib/kubelet/pods/e16d795b-bf24-4faa-81b5-d7b1d08bb5a4/volumes" Nov 25 10:08:36 crc kubenswrapper[4926]: I1125 10:08:36.074527 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Nov 25 10:08:36 crc kubenswrapper[4926]: W1125 10:08:36.077104 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff3b4281_7815_4363_91b6_fabe18bfdb70.slice/crio-ba03b10dec83d1e5f4365f759ceb1be0b1a668868033041d1f3f08c059d0ce2b WatchSource:0}: Error finding container ba03b10dec83d1e5f4365f759ceb1be0b1a668868033041d1f3f08c059d0ce2b: Status 404 returned error can't find the container with id ba03b10dec83d1e5f4365f759ceb1be0b1a668868033041d1f3f08c059d0ce2b Nov 25 10:08:36 crc kubenswrapper[4926]: I1125 10:08:36.954493 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ff3b4281-7815-4363-91b6-fabe18bfdb70","Type":"ContainerStarted","Data":"aa8bbd401cb311ac28e304f53c5e83376da2f4d2c430710fe0615fba87ff1cdd"} Nov 25 10:08:36 crc kubenswrapper[4926]: I1125 10:08:36.954839 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ff3b4281-7815-4363-91b6-fabe18bfdb70","Type":"ContainerStarted","Data":"ba03b10dec83d1e5f4365f759ceb1be0b1a668868033041d1f3f08c059d0ce2b"} Nov 25 10:08:36 crc kubenswrapper[4926]: I1125 10:08:36.978592 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.978566647 podStartE2EDuration="1.978566647s" podCreationTimestamp="2025-11-25 10:08:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:36.967950393 +0000 UTC m=+1213.651906840" watchObservedRunningTime="2025-11-25 10:08:36.978566647 +0000 UTC m=+1213.662523084" Nov 25 10:08:37 crc kubenswrapper[4926]: I1125 10:08:37.231310 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:08:37 crc kubenswrapper[4926]: I1125 10:08:37.232085 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:08:37 crc kubenswrapper[4926]: I1125 10:08:37.232206 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:08:37 crc kubenswrapper[4926]: I1125 10:08:37.236588 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:08:37 crc kubenswrapper[4926]: I1125 10:08:37.962460 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:08:37 crc kubenswrapper[4926]: I1125 10:08:37.965853 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.184964 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-sbmk8"] Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.187124 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.204012 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-sbmk8"] Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.388033 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bd5xf\" (UniqueName: \"kubernetes.io/projected/1a66f16d-89fd-4885-bc00-87f73e6705e8-kube-api-access-bd5xf\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.388154 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.388185 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.388215 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-config\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.388455 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.491252 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.491366 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bd5xf\" (UniqueName: \"kubernetes.io/projected/1a66f16d-89fd-4885-bc00-87f73e6705e8-kube-api-access-bd5xf\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.491449 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.491478 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.491509 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-config\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.492694 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-config\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.492699 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-nb\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.492817 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-sb\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.492823 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-dns-svc\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.510393 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bd5xf\" (UniqueName: \"kubernetes.io/projected/1a66f16d-89fd-4885-bc00-87f73e6705e8-kube-api-access-bd5xf\") pod \"dnsmasq-dns-68d4b6d797-sbmk8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:38 crc kubenswrapper[4926]: I1125 10:08:38.521769 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:39 crc kubenswrapper[4926]: I1125 10:08:39.005877 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-sbmk8"] Nov 25 10:08:39 crc kubenswrapper[4926]: W1125 10:08:39.022283 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a66f16d_89fd_4885_bc00_87f73e6705e8.slice/crio-84647ade3f46e4e1ca2fd993772b85895e1f062b43a0a112473e4e07d41044c4 WatchSource:0}: Error finding container 84647ade3f46e4e1ca2fd993772b85895e1f062b43a0a112473e4e07d41044c4: Status 404 returned error can't find the container with id 84647ade3f46e4e1ca2fd993772b85895e1f062b43a0a112473e4e07d41044c4 Nov 25 10:08:39 crc kubenswrapper[4926]: I1125 10:08:39.979692 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerID="dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b" exitCode=0 Nov 25 10:08:39 crc kubenswrapper[4926]: I1125 10:08:39.979812 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" event={"ID":"1a66f16d-89fd-4885-bc00-87f73e6705e8","Type":"ContainerDied","Data":"dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b"} Nov 25 10:08:39 crc kubenswrapper[4926]: I1125 10:08:39.981114 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" event={"ID":"1a66f16d-89fd-4885-bc00-87f73e6705e8","Type":"ContainerStarted","Data":"84647ade3f46e4e1ca2fd993772b85895e1f062b43a0a112473e4e07d41044c4"} Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.412724 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.413035 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-central-agent" containerID="cri-o://05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4" gracePeriod=30 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.413153 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="proxy-httpd" containerID="cri-o://ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0" gracePeriod=30 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.413201 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="sg-core" containerID="cri-o://3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb" gracePeriod=30 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.413235 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-notification-agent" containerID="cri-o://78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d" gracePeriod=30 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.659639 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.673316 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.990725 4926 generic.go:334] "Generic (PLEG): container finished" podID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerID="ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0" exitCode=0 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.990761 4926 generic.go:334] "Generic (PLEG): container finished" podID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerID="3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb" exitCode=2 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.990772 4926 generic.go:334] "Generic (PLEG): container finished" podID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerID="05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4" exitCode=0 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.990815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerDied","Data":"ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0"} Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.990850 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerDied","Data":"3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb"} Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.990870 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerDied","Data":"05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4"} Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.995531 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-log" containerID="cri-o://5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d" gracePeriod=30 Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.996803 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" event={"ID":"1a66f16d-89fd-4885-bc00-87f73e6705e8","Type":"ContainerStarted","Data":"7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401"} Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.996850 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:40 crc kubenswrapper[4926]: I1125 10:08:40.997366 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-api" containerID="cri-o://c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322" gracePeriod=30 Nov 25 10:08:41 crc kubenswrapper[4926]: I1125 10:08:41.034456 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" podStartSLOduration=3.03443191 podStartE2EDuration="3.03443191s" podCreationTimestamp="2025-11-25 10:08:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:41.024998505 +0000 UTC m=+1217.708954932" watchObservedRunningTime="2025-11-25 10:08:41.03443191 +0000 UTC m=+1217.718388347" Nov 25 10:08:42 crc kubenswrapper[4926]: I1125 10:08:42.004616 4926 generic.go:334] "Generic (PLEG): container finished" podID="2822671e-433a-4983-81cf-c315f9a87184" containerID="5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d" exitCode=143 Nov 25 10:08:42 crc kubenswrapper[4926]: I1125 10:08:42.004696 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2822671e-433a-4983-81cf-c315f9a87184","Type":"ContainerDied","Data":"5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d"} Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.580276 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.695475 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-config-data\") pod \"2822671e-433a-4983-81cf-c315f9a87184\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.695540 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-combined-ca-bundle\") pod \"2822671e-433a-4983-81cf-c315f9a87184\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.695679 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2822671e-433a-4983-81cf-c315f9a87184-logs\") pod \"2822671e-433a-4983-81cf-c315f9a87184\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.695782 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztmz7\" (UniqueName: \"kubernetes.io/projected/2822671e-433a-4983-81cf-c315f9a87184-kube-api-access-ztmz7\") pod \"2822671e-433a-4983-81cf-c315f9a87184\" (UID: \"2822671e-433a-4983-81cf-c315f9a87184\") " Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.697083 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2822671e-433a-4983-81cf-c315f9a87184-logs" (OuterVolumeSpecName: "logs") pod "2822671e-433a-4983-81cf-c315f9a87184" (UID: "2822671e-433a-4983-81cf-c315f9a87184"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.712850 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2822671e-433a-4983-81cf-c315f9a87184-kube-api-access-ztmz7" (OuterVolumeSpecName: "kube-api-access-ztmz7") pod "2822671e-433a-4983-81cf-c315f9a87184" (UID: "2822671e-433a-4983-81cf-c315f9a87184"). InnerVolumeSpecName "kube-api-access-ztmz7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.738865 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-config-data" (OuterVolumeSpecName: "config-data") pod "2822671e-433a-4983-81cf-c315f9a87184" (UID: "2822671e-433a-4983-81cf-c315f9a87184"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.746658 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2822671e-433a-4983-81cf-c315f9a87184" (UID: "2822671e-433a-4983-81cf-c315f9a87184"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.797466 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2822671e-433a-4983-81cf-c315f9a87184-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.797500 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztmz7\" (UniqueName: \"kubernetes.io/projected/2822671e-433a-4983-81cf-c315f9a87184-kube-api-access-ztmz7\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.797516 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:44 crc kubenswrapper[4926]: I1125 10:08:44.797527 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2822671e-433a-4983-81cf-c315f9a87184-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.036609 4926 generic.go:334] "Generic (PLEG): container finished" podID="2822671e-433a-4983-81cf-c315f9a87184" containerID="c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322" exitCode=0 Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.036683 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2822671e-433a-4983-81cf-c315f9a87184","Type":"ContainerDied","Data":"c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322"} Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.036708 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2822671e-433a-4983-81cf-c315f9a87184","Type":"ContainerDied","Data":"c41baa2764ab8f086a7a2d3ec812e909f856ebe7590af68182598392f283f379"} Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.036724 4926 scope.go:117] "RemoveContainer" containerID="c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.036832 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.076252 4926 scope.go:117] "RemoveContainer" containerID="5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.077038 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.091576 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.109690 4926 scope.go:117] "RemoveContainer" containerID="c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322" Nov 25 10:08:45 crc kubenswrapper[4926]: E1125 10:08:45.110152 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322\": container with ID starting with c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322 not found: ID does not exist" containerID="c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.110192 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322"} err="failed to get container status \"c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322\": rpc error: code = NotFound desc = could not find container \"c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322\": container with ID starting with c8fcd625778efaa0e2549d9e8b91ce3ac775b1274fd6871f5db8b1512ba08322 not found: ID does not exist" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.110220 4926 scope.go:117] "RemoveContainer" containerID="5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d" Nov 25 10:08:45 crc kubenswrapper[4926]: E1125 10:08:45.110515 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d\": container with ID starting with 5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d not found: ID does not exist" containerID="5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.110546 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d"} err="failed to get container status \"5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d\": rpc error: code = NotFound desc = could not find container \"5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d\": container with ID starting with 5374f5c43dabea76cf8b83b36eedbbdfda0411bf1c103754494bf9513f99fb4d not found: ID does not exist" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.113208 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:45 crc kubenswrapper[4926]: E1125 10:08:45.113576 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-log" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.113594 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-log" Nov 25 10:08:45 crc kubenswrapper[4926]: E1125 10:08:45.113648 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-api" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.113656 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-api" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.113878 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-log" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.113910 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2822671e-433a-4983-81cf-c315f9a87184" containerName="nova-api-api" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.115077 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.117837 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.118095 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.118239 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.143272 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.305327 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.305365 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.305524 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ea4b-a014-4954-bbfc-becf710f80ee-logs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.305559 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.305602 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-config-data\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.305629 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6479k\" (UniqueName: \"kubernetes.io/projected/3dc1ea4b-a014-4954-bbfc-becf710f80ee-kube-api-access-6479k\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.407572 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.407624 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ea4b-a014-4954-bbfc-becf710f80ee-logs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.407674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-config-data\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.407703 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6479k\" (UniqueName: \"kubernetes.io/projected/3dc1ea4b-a014-4954-bbfc-becf710f80ee-kube-api-access-6479k\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.407753 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.407781 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.408112 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ea4b-a014-4954-bbfc-becf710f80ee-logs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.413382 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-config-data\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.413599 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-public-tls-certs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.417258 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.418368 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.427975 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6479k\" (UniqueName: \"kubernetes.io/projected/3dc1ea4b-a014-4954-bbfc-becf710f80ee-kube-api-access-6479k\") pod \"nova-api-0\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.449560 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.660603 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.679362 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:45 crc kubenswrapper[4926]: I1125 10:08:45.930192 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.024487 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2822671e-433a-4983-81cf-c315f9a87184" path="/var/lib/kubelet/pods/2822671e-433a-4983-81cf-c315f9a87184/volumes" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.047574 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dc1ea4b-a014-4954-bbfc-becf710f80ee","Type":"ContainerStarted","Data":"ba155c951f5207140eae31eada63a3238f23a741f3f6e1201236eca919dd0c75"} Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.070202 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.213764 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-mzsx6"] Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.215419 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.219868 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.220261 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.229033 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mzsx6"] Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.327284 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-config-data\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.327660 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7rm9\" (UniqueName: \"kubernetes.io/projected/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-kube-api-access-r7rm9\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.327741 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.327915 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-scripts\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.430191 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-config-data\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.430583 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7rm9\" (UniqueName: \"kubernetes.io/projected/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-kube-api-access-r7rm9\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.430654 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.430796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-scripts\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.434624 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.435113 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-config-data\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.436337 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-scripts\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.450417 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7rm9\" (UniqueName: \"kubernetes.io/projected/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-kube-api-access-r7rm9\") pod \"nova-cell1-cell-mapping-mzsx6\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:46 crc kubenswrapper[4926]: I1125 10:08:46.552382 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:47 crc kubenswrapper[4926]: I1125 10:08:47.061909 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dc1ea4b-a014-4954-bbfc-becf710f80ee","Type":"ContainerStarted","Data":"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf"} Nov 25 10:08:47 crc kubenswrapper[4926]: I1125 10:08:47.062194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dc1ea4b-a014-4954-bbfc-becf710f80ee","Type":"ContainerStarted","Data":"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f"} Nov 25 10:08:47 crc kubenswrapper[4926]: I1125 10:08:47.138420 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.13839886 podStartE2EDuration="2.13839886s" podCreationTimestamp="2025-11-25 10:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:47.093828389 +0000 UTC m=+1223.777784826" watchObservedRunningTime="2025-11-25 10:08:47.13839886 +0000 UTC m=+1223.822355297" Nov 25 10:08:47 crc kubenswrapper[4926]: I1125 10:08:47.149898 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-mzsx6"] Nov 25 10:08:48 crc kubenswrapper[4926]: I1125 10:08:48.083342 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mzsx6" event={"ID":"ec51083f-ad2b-4b95-ba9f-de0a8db553fc","Type":"ContainerStarted","Data":"9b382af683c76723ffeebe2758d190df26f88b17ebe44ced6b91fc0841029d6f"} Nov 25 10:08:48 crc kubenswrapper[4926]: I1125 10:08:48.084004 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mzsx6" event={"ID":"ec51083f-ad2b-4b95-ba9f-de0a8db553fc","Type":"ContainerStarted","Data":"b29a86ee68f211689fe70b9517e67dcafb683a1878cfe5c9d1e338e001bf3d30"} Nov 25 10:08:48 crc kubenswrapper[4926]: I1125 10:08:48.106051 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-mzsx6" podStartSLOduration=2.10603282 podStartE2EDuration="2.10603282s" podCreationTimestamp="2025-11-25 10:08:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:48.104094782 +0000 UTC m=+1224.788051259" watchObservedRunningTime="2025-11-25 10:08:48.10603282 +0000 UTC m=+1224.789989267" Nov 25 10:08:48 crc kubenswrapper[4926]: I1125 10:08:48.523897 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:08:48 crc kubenswrapper[4926]: I1125 10:08:48.597730 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-ft7kt"] Nov 25 10:08:48 crc kubenswrapper[4926]: I1125 10:08:48.598087 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerName="dnsmasq-dns" containerID="cri-o://acdc94c2d20d7b8c6619f73c327ef238d106843c43f5bb29fe9ba111c385f3f3" gracePeriod=10 Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.091399 4926 generic.go:334] "Generic (PLEG): container finished" podID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerID="acdc94c2d20d7b8c6619f73c327ef238d106843c43f5bb29fe9ba111c385f3f3" exitCode=0 Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.091591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" event={"ID":"b86d5f2e-290b-4856-b454-e86800fdd30e","Type":"ContainerDied","Data":"acdc94c2d20d7b8c6619f73c327ef238d106843c43f5bb29fe9ba111c385f3f3"} Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.091755 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" event={"ID":"b86d5f2e-290b-4856-b454-e86800fdd30e","Type":"ContainerDied","Data":"4b53ed51f623557b6350c820e2b89f8e90ce04b29ffc19444a8b6da2342f519f"} Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.091774 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b53ed51f623557b6350c820e2b89f8e90ce04b29ffc19444a8b6da2342f519f" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.124038 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.321488 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-nb\") pod \"b86d5f2e-290b-4856-b454-e86800fdd30e\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.321639 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-config\") pod \"b86d5f2e-290b-4856-b454-e86800fdd30e\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.321699 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c7m2\" (UniqueName: \"kubernetes.io/projected/b86d5f2e-290b-4856-b454-e86800fdd30e-kube-api-access-7c7m2\") pod \"b86d5f2e-290b-4856-b454-e86800fdd30e\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.321748 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-dns-svc\") pod \"b86d5f2e-290b-4856-b454-e86800fdd30e\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.321802 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-sb\") pod \"b86d5f2e-290b-4856-b454-e86800fdd30e\" (UID: \"b86d5f2e-290b-4856-b454-e86800fdd30e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.347446 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b86d5f2e-290b-4856-b454-e86800fdd30e-kube-api-access-7c7m2" (OuterVolumeSpecName: "kube-api-access-7c7m2") pod "b86d5f2e-290b-4856-b454-e86800fdd30e" (UID: "b86d5f2e-290b-4856-b454-e86800fdd30e"). InnerVolumeSpecName "kube-api-access-7c7m2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.404482 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b86d5f2e-290b-4856-b454-e86800fdd30e" (UID: "b86d5f2e-290b-4856-b454-e86800fdd30e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.405164 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b86d5f2e-290b-4856-b454-e86800fdd30e" (UID: "b86d5f2e-290b-4856-b454-e86800fdd30e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.411204 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b86d5f2e-290b-4856-b454-e86800fdd30e" (UID: "b86d5f2e-290b-4856-b454-e86800fdd30e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.426433 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.426466 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c7m2\" (UniqueName: \"kubernetes.io/projected/b86d5f2e-290b-4856-b454-e86800fdd30e-kube-api-access-7c7m2\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.426507 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.426521 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.440373 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-config" (OuterVolumeSpecName: "config") pod "b86d5f2e-290b-4856-b454-e86800fdd30e" (UID: "b86d5f2e-290b-4856-b454-e86800fdd30e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.527855 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b86d5f2e-290b-4856-b454-e86800fdd30e-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.804409 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.935675 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-sg-core-conf-yaml\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.935754 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-ceilometer-tls-certs\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.935902 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-scripts\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.935965 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-config-data\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.936012 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-combined-ca-bundle\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.936057 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-run-httpd\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.936133 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-log-httpd\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.936685 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.936888 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.936201 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpbfd\" (UniqueName: \"kubernetes.io/projected/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-kube-api-access-qpbfd\") pod \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\" (UID: \"3bc01e08-29e2-4a36-b06f-8dac6c33de8e\") " Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.940924 4926 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-log-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.940973 4926 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-run-httpd\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.949127 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-scripts" (OuterVolumeSpecName: "scripts") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.965198 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-kube-api-access-qpbfd" (OuterVolumeSpecName: "kube-api-access-qpbfd") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "kube-api-access-qpbfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:49 crc kubenswrapper[4926]: I1125 10:08:49.984593 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.022054 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.044089 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.044125 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpbfd\" (UniqueName: \"kubernetes.io/projected/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-kube-api-access-qpbfd\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.044140 4926 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.044153 4926 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.063241 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.089629 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-config-data" (OuterVolumeSpecName: "config-data") pod "3bc01e08-29e2-4a36-b06f-8dac6c33de8e" (UID: "3bc01e08-29e2-4a36-b06f-8dac6c33de8e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.102232 4926 generic.go:334] "Generic (PLEG): container finished" podID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerID="78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d" exitCode=0 Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.104234 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b8cf6657-ft7kt" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.103063 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.147168 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerDied","Data":"78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d"} Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.147408 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3bc01e08-29e2-4a36-b06f-8dac6c33de8e","Type":"ContainerDied","Data":"925fc809ae48226d0d5c0c36241634ccfba2646ba2321ba7d9edaf73788d7000"} Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.147524 4926 scope.go:117] "RemoveContainer" containerID="ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.150646 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.150825 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3bc01e08-29e2-4a36-b06f-8dac6c33de8e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.215994 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.221082 4926 scope.go:117] "RemoveContainer" containerID="3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.226339 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.235025 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-ft7kt"] Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.246083 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b8cf6657-ft7kt"] Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254171 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.254485 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-notification-agent" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254501 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-notification-agent" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.254529 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerName="init" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254535 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerName="init" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.254562 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerName="dnsmasq-dns" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254573 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerName="dnsmasq-dns" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.254582 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-central-agent" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254590 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-central-agent" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.254602 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="sg-core" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254607 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="sg-core" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.254622 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="proxy-httpd" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254628 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="proxy-httpd" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254769 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="proxy-httpd" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254783 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" containerName="dnsmasq-dns" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254797 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="sg-core" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254808 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-central-agent" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.254820 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" containerName="ceilometer-notification-agent" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.256338 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.260529 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.268423 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.268662 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.268867 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.269031 4926 scope.go:117] "RemoveContainer" containerID="78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.299410 4926 scope.go:117] "RemoveContainer" containerID="05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.331956 4926 scope.go:117] "RemoveContainer" containerID="ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.332330 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0\": container with ID starting with ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0 not found: ID does not exist" containerID="ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.332378 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0"} err="failed to get container status \"ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0\": rpc error: code = NotFound desc = could not find container \"ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0\": container with ID starting with ba2b0e963bf3d220078d21c84a8064e27841b7b5158535f2b50bf263294481c0 not found: ID does not exist" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.332404 4926 scope.go:117] "RemoveContainer" containerID="3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.332749 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb\": container with ID starting with 3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb not found: ID does not exist" containerID="3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.332782 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb"} err="failed to get container status \"3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb\": rpc error: code = NotFound desc = could not find container \"3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb\": container with ID starting with 3008470b02564fcf9bdc3b8af4fdd27c45711e32bac04538c777a8be9614cedb not found: ID does not exist" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.332804 4926 scope.go:117] "RemoveContainer" containerID="78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.333022 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d\": container with ID starting with 78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d not found: ID does not exist" containerID="78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.333038 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d"} err="failed to get container status \"78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d\": rpc error: code = NotFound desc = could not find container \"78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d\": container with ID starting with 78d03b66c4c6ee0cf09ccf8acd0cb4e6600540a209f05b3dea3890051e50be5d not found: ID does not exist" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.333050 4926 scope.go:117] "RemoveContainer" containerID="05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4" Nov 25 10:08:50 crc kubenswrapper[4926]: E1125 10:08:50.333641 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4\": container with ID starting with 05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4 not found: ID does not exist" containerID="05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.333667 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4"} err="failed to get container status \"05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4\": rpc error: code = NotFound desc = could not find container \"05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4\": container with ID starting with 05186b3fad8f0c36441eff46c5bcf25e5664a02bf0f326736d440582c0f7d5e4 not found: ID does not exist" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.355687 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.355734 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.355758 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dt7k\" (UniqueName: \"kubernetes.io/projected/0a649a79-6eff-402f-8ea0-59b8649155e3-kube-api-access-5dt7k\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.355827 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a649a79-6eff-402f-8ea0-59b8649155e3-run-httpd\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.355867 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.355966 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-scripts\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.356011 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-config-data\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.356084 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a649a79-6eff-402f-8ea0-59b8649155e3-log-httpd\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456618 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456659 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456682 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dt7k\" (UniqueName: \"kubernetes.io/projected/0a649a79-6eff-402f-8ea0-59b8649155e3-kube-api-access-5dt7k\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456708 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a649a79-6eff-402f-8ea0-59b8649155e3-run-httpd\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456736 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456764 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-scripts\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456784 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-config-data\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.456827 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a649a79-6eff-402f-8ea0-59b8649155e3-log-httpd\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.457244 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a649a79-6eff-402f-8ea0-59b8649155e3-log-httpd\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.457834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0a649a79-6eff-402f-8ea0-59b8649155e3-run-httpd\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.461701 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.462076 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-scripts\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.462297 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.462538 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-config-data\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.462985 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0a649a79-6eff-402f-8ea0-59b8649155e3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.474100 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dt7k\" (UniqueName: \"kubernetes.io/projected/0a649a79-6eff-402f-8ea0-59b8649155e3-kube-api-access-5dt7k\") pod \"ceilometer-0\" (UID: \"0a649a79-6eff-402f-8ea0-59b8649155e3\") " pod="openstack/ceilometer-0" Nov 25 10:08:50 crc kubenswrapper[4926]: I1125 10:08:50.580238 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Nov 25 10:08:51 crc kubenswrapper[4926]: I1125 10:08:51.070470 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Nov 25 10:08:51 crc kubenswrapper[4926]: W1125 10:08:51.074134 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0a649a79_6eff_402f_8ea0_59b8649155e3.slice/crio-eae7793737423d3f53d3421d9167720f38b34ff9f002551592e987a0f1a7badd WatchSource:0}: Error finding container eae7793737423d3f53d3421d9167720f38b34ff9f002551592e987a0f1a7badd: Status 404 returned error can't find the container with id eae7793737423d3f53d3421d9167720f38b34ff9f002551592e987a0f1a7badd Nov 25 10:08:51 crc kubenswrapper[4926]: I1125 10:08:51.113145 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a649a79-6eff-402f-8ea0-59b8649155e3","Type":"ContainerStarted","Data":"eae7793737423d3f53d3421d9167720f38b34ff9f002551592e987a0f1a7badd"} Nov 25 10:08:52 crc kubenswrapper[4926]: I1125 10:08:52.022617 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3bc01e08-29e2-4a36-b06f-8dac6c33de8e" path="/var/lib/kubelet/pods/3bc01e08-29e2-4a36-b06f-8dac6c33de8e/volumes" Nov 25 10:08:52 crc kubenswrapper[4926]: I1125 10:08:52.023884 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b86d5f2e-290b-4856-b454-e86800fdd30e" path="/var/lib/kubelet/pods/b86d5f2e-290b-4856-b454-e86800fdd30e/volumes" Nov 25 10:08:52 crc kubenswrapper[4926]: I1125 10:08:52.122301 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a649a79-6eff-402f-8ea0-59b8649155e3","Type":"ContainerStarted","Data":"ed49de58170b54b08a5df860c428918ba0b8eb496651446179b60f1d66994f19"} Nov 25 10:08:53 crc kubenswrapper[4926]: I1125 10:08:53.161600 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a649a79-6eff-402f-8ea0-59b8649155e3","Type":"ContainerStarted","Data":"f29f0ed0be372a20ce26318cf355da73009d14032a11db78fa115b27d6577114"} Nov 25 10:08:53 crc kubenswrapper[4926]: I1125 10:08:53.164577 4926 generic.go:334] "Generic (PLEG): container finished" podID="ec51083f-ad2b-4b95-ba9f-de0a8db553fc" containerID="9b382af683c76723ffeebe2758d190df26f88b17ebe44ced6b91fc0841029d6f" exitCode=0 Nov 25 10:08:53 crc kubenswrapper[4926]: I1125 10:08:53.164627 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mzsx6" event={"ID":"ec51083f-ad2b-4b95-ba9f-de0a8db553fc","Type":"ContainerDied","Data":"9b382af683c76723ffeebe2758d190df26f88b17ebe44ced6b91fc0841029d6f"} Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.173868 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a649a79-6eff-402f-8ea0-59b8649155e3","Type":"ContainerStarted","Data":"76c06a45db7483f75d3d9f26356ff0449f802e4dc901376aab32bae5ce323bf7"} Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.594418 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.731714 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-config-data\") pod \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.731775 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7rm9\" (UniqueName: \"kubernetes.io/projected/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-kube-api-access-r7rm9\") pod \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.731793 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-scripts\") pod \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.731833 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-combined-ca-bundle\") pod \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\" (UID: \"ec51083f-ad2b-4b95-ba9f-de0a8db553fc\") " Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.736653 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-kube-api-access-r7rm9" (OuterVolumeSpecName: "kube-api-access-r7rm9") pod "ec51083f-ad2b-4b95-ba9f-de0a8db553fc" (UID: "ec51083f-ad2b-4b95-ba9f-de0a8db553fc"). InnerVolumeSpecName "kube-api-access-r7rm9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.736866 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-scripts" (OuterVolumeSpecName: "scripts") pod "ec51083f-ad2b-4b95-ba9f-de0a8db553fc" (UID: "ec51083f-ad2b-4b95-ba9f-de0a8db553fc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.754528 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ec51083f-ad2b-4b95-ba9f-de0a8db553fc" (UID: "ec51083f-ad2b-4b95-ba9f-de0a8db553fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.765347 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-config-data" (OuterVolumeSpecName: "config-data") pod "ec51083f-ad2b-4b95-ba9f-de0a8db553fc" (UID: "ec51083f-ad2b-4b95-ba9f-de0a8db553fc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.834335 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.834370 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7rm9\" (UniqueName: \"kubernetes.io/projected/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-kube-api-access-r7rm9\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.834382 4926 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-scripts\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:54 crc kubenswrapper[4926]: I1125 10:08:54.834391 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ec51083f-ad2b-4b95-ba9f-de0a8db553fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.183903 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0a649a79-6eff-402f-8ea0-59b8649155e3","Type":"ContainerStarted","Data":"82bf1b4715d766555732db8111bfe2d911a0ef477e604119be1cace1fd882ffc"} Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.186642 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.187967 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-mzsx6" event={"ID":"ec51083f-ad2b-4b95-ba9f-de0a8db553fc","Type":"ContainerDied","Data":"b29a86ee68f211689fe70b9517e67dcafb683a1878cfe5c9d1e338e001bf3d30"} Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.188000 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b29a86ee68f211689fe70b9517e67dcafb683a1878cfe5c9d1e338e001bf3d30" Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.188097 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-mzsx6" Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.223369 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.317922856 podStartE2EDuration="5.22334722s" podCreationTimestamp="2025-11-25 10:08:50 +0000 UTC" firstStartedPulling="2025-11-25 10:08:51.076597691 +0000 UTC m=+1227.760554128" lastFinishedPulling="2025-11-25 10:08:54.982022065 +0000 UTC m=+1231.665978492" observedRunningTime="2025-11-25 10:08:55.212369466 +0000 UTC m=+1231.896325893" watchObservedRunningTime="2025-11-25 10:08:55.22334722 +0000 UTC m=+1231.907303647" Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.354577 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.355107 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-log" containerID="cri-o://09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f" gracePeriod=30 Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.355161 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-api" containerID="cri-o://b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf" gracePeriod=30 Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.372888 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.373486 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="406430ac-1468-495b-a177-52e52f5a0405" containerName="nova-scheduler-scheduler" containerID="cri-o://f9ad0b36a3f22b70a06652dc863be694e8b4662461d428a18e1ed6a19f3c6ed8" gracePeriod=30 Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.403058 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.403583 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-log" containerID="cri-o://61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9" gracePeriod=30 Nov 25 10:08:55 crc kubenswrapper[4926]: I1125 10:08:55.403661 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-metadata" containerID="cri-o://3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711" gracePeriod=30 Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.037591 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.156744 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-combined-ca-bundle\") pod \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.156801 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6479k\" (UniqueName: \"kubernetes.io/projected/3dc1ea4b-a014-4954-bbfc-becf710f80ee-kube-api-access-6479k\") pod \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.156870 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-public-tls-certs\") pod \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.157009 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-config-data\") pod \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.157053 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-internal-tls-certs\") pod \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.157085 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ea4b-a014-4954-bbfc-becf710f80ee-logs\") pod \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\" (UID: \"3dc1ea4b-a014-4954-bbfc-becf710f80ee\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.157680 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dc1ea4b-a014-4954-bbfc-becf710f80ee-logs" (OuterVolumeSpecName: "logs") pod "3dc1ea4b-a014-4954-bbfc-becf710f80ee" (UID: "3dc1ea4b-a014-4954-bbfc-becf710f80ee"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.166192 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dc1ea4b-a014-4954-bbfc-becf710f80ee-kube-api-access-6479k" (OuterVolumeSpecName: "kube-api-access-6479k") pod "3dc1ea4b-a014-4954-bbfc-becf710f80ee" (UID: "3dc1ea4b-a014-4954-bbfc-becf710f80ee"). InnerVolumeSpecName "kube-api-access-6479k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.206452 4926 generic.go:334] "Generic (PLEG): container finished" podID="406430ac-1468-495b-a177-52e52f5a0405" containerID="f9ad0b36a3f22b70a06652dc863be694e8b4662461d428a18e1ed6a19f3c6ed8" exitCode=0 Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.206774 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406430ac-1468-495b-a177-52e52f5a0405","Type":"ContainerDied","Data":"f9ad0b36a3f22b70a06652dc863be694e8b4662461d428a18e1ed6a19f3c6ed8"} Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212228 4926 generic.go:334] "Generic (PLEG): container finished" podID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerID="b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf" exitCode=0 Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212257 4926 generic.go:334] "Generic (PLEG): container finished" podID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerID="09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f" exitCode=143 Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212297 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dc1ea4b-a014-4954-bbfc-becf710f80ee","Type":"ContainerDied","Data":"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf"} Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212325 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dc1ea4b-a014-4954-bbfc-becf710f80ee","Type":"ContainerDied","Data":"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f"} Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3dc1ea4b-a014-4954-bbfc-becf710f80ee","Type":"ContainerDied","Data":"ba155c951f5207140eae31eada63a3238f23a741f3f6e1201236eca919dd0c75"} Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212357 4926 scope.go:117] "RemoveContainer" containerID="b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212500 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.212596 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3dc1ea4b-a014-4954-bbfc-becf710f80ee" (UID: "3dc1ea4b-a014-4954-bbfc-becf710f80ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.222400 4926 generic.go:334] "Generic (PLEG): container finished" podID="1760dcfa-c766-4bf2-b054-223863000927" containerID="61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9" exitCode=143 Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.223359 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1760dcfa-c766-4bf2-b054-223863000927","Type":"ContainerDied","Data":"61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9"} Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.260584 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.260619 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6479k\" (UniqueName: \"kubernetes.io/projected/3dc1ea4b-a014-4954-bbfc-becf710f80ee-kube-api-access-6479k\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.260635 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3dc1ea4b-a014-4954-bbfc-becf710f80ee-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.262418 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-config-data" (OuterVolumeSpecName: "config-data") pod "3dc1ea4b-a014-4954-bbfc-becf710f80ee" (UID: "3dc1ea4b-a014-4954-bbfc-becf710f80ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.262440 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "3dc1ea4b-a014-4954-bbfc-becf710f80ee" (UID: "3dc1ea4b-a014-4954-bbfc-becf710f80ee"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.273077 4926 scope.go:117] "RemoveContainer" containerID="09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.273069 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "3dc1ea4b-a014-4954-bbfc-becf710f80ee" (UID: "3dc1ea4b-a014-4954-bbfc-becf710f80ee"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.297795 4926 scope.go:117] "RemoveContainer" containerID="b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf" Nov 25 10:08:56 crc kubenswrapper[4926]: E1125 10:08:56.303102 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf\": container with ID starting with b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf not found: ID does not exist" containerID="b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.303169 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf"} err="failed to get container status \"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf\": rpc error: code = NotFound desc = could not find container \"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf\": container with ID starting with b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf not found: ID does not exist" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.303205 4926 scope.go:117] "RemoveContainer" containerID="09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f" Nov 25 10:08:56 crc kubenswrapper[4926]: E1125 10:08:56.303801 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f\": container with ID starting with 09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f not found: ID does not exist" containerID="09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.303831 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f"} err="failed to get container status \"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f\": rpc error: code = NotFound desc = could not find container \"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f\": container with ID starting with 09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f not found: ID does not exist" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.303846 4926 scope.go:117] "RemoveContainer" containerID="b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.304333 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf"} err="failed to get container status \"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf\": rpc error: code = NotFound desc = could not find container \"b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf\": container with ID starting with b18abc822b0c9fea25aa5012b3c066522ae4368b41fc02516bb8f251f99debbf not found: ID does not exist" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.304357 4926 scope.go:117] "RemoveContainer" containerID="09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.304725 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f"} err="failed to get container status \"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f\": rpc error: code = NotFound desc = could not find container \"09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f\": container with ID starting with 09f778741343edcaf89bd86d354bce5050c52d1b3dd799cf50c71f5e70b0f22f not found: ID does not exist" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.316727 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.362113 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.362145 4926 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.362155 4926 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3dc1ea4b-a014-4954-bbfc-becf710f80ee-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.463475 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kb5s\" (UniqueName: \"kubernetes.io/projected/406430ac-1468-495b-a177-52e52f5a0405-kube-api-access-4kb5s\") pod \"406430ac-1468-495b-a177-52e52f5a0405\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.463584 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-combined-ca-bundle\") pod \"406430ac-1468-495b-a177-52e52f5a0405\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.463680 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-config-data\") pod \"406430ac-1468-495b-a177-52e52f5a0405\" (UID: \"406430ac-1468-495b-a177-52e52f5a0405\") " Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.468111 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/406430ac-1468-495b-a177-52e52f5a0405-kube-api-access-4kb5s" (OuterVolumeSpecName: "kube-api-access-4kb5s") pod "406430ac-1468-495b-a177-52e52f5a0405" (UID: "406430ac-1468-495b-a177-52e52f5a0405"). InnerVolumeSpecName "kube-api-access-4kb5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.489877 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-config-data" (OuterVolumeSpecName: "config-data") pod "406430ac-1468-495b-a177-52e52f5a0405" (UID: "406430ac-1468-495b-a177-52e52f5a0405"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.494276 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "406430ac-1468-495b-a177-52e52f5a0405" (UID: "406430ac-1468-495b-a177-52e52f5a0405"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.566682 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kb5s\" (UniqueName: \"kubernetes.io/projected/406430ac-1468-495b-a177-52e52f5a0405-kube-api-access-4kb5s\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.566711 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.566721 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/406430ac-1468-495b-a177-52e52f5a0405-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.587741 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.594821 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.611914 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:56 crc kubenswrapper[4926]: E1125 10:08:56.612319 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-log" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612338 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-log" Nov 25 10:08:56 crc kubenswrapper[4926]: E1125 10:08:56.612360 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="406430ac-1468-495b-a177-52e52f5a0405" containerName="nova-scheduler-scheduler" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612367 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="406430ac-1468-495b-a177-52e52f5a0405" containerName="nova-scheduler-scheduler" Nov 25 10:08:56 crc kubenswrapper[4926]: E1125 10:08:56.612381 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec51083f-ad2b-4b95-ba9f-de0a8db553fc" containerName="nova-manage" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612389 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec51083f-ad2b-4b95-ba9f-de0a8db553fc" containerName="nova-manage" Nov 25 10:08:56 crc kubenswrapper[4926]: E1125 10:08:56.612404 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-api" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612410 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-api" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612557 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec51083f-ad2b-4b95-ba9f-de0a8db553fc" containerName="nova-manage" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612576 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-log" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612585 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="406430ac-1468-495b-a177-52e52f5a0405" containerName="nova-scheduler-scheduler" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.612597 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" containerName="nova-api-api" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.613524 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.622038 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.622204 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.622343 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.627460 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.769455 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.769502 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.769533 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.769682 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-config-data\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.769749 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cnln\" (UniqueName: \"kubernetes.io/projected/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-kube-api-access-2cnln\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.769822 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-logs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.871446 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cnln\" (UniqueName: \"kubernetes.io/projected/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-kube-api-access-2cnln\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.871810 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-logs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.872017 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.872043 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.872079 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.872119 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-config-data\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.872268 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-logs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.877430 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.878618 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-public-tls-certs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.886202 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-internal-tls-certs\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.889140 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-config-data\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.894834 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cnln\" (UniqueName: \"kubernetes.io/projected/de6e1390-0172-4a3f-b6cb-5abb2a5633b2-kube-api-access-2cnln\") pod \"nova-api-0\" (UID: \"de6e1390-0172-4a3f-b6cb-5abb2a5633b2\") " pod="openstack/nova-api-0" Nov 25 10:08:56 crc kubenswrapper[4926]: I1125 10:08:56.930053 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.233336 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"406430ac-1468-495b-a177-52e52f5a0405","Type":"ContainerDied","Data":"c4e38acc019ca436bf810ddea2c5901d6e90f55cfc78286b774841f43eebc03d"} Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.233383 4926 scope.go:117] "RemoveContainer" containerID="f9ad0b36a3f22b70a06652dc863be694e8b4662461d428a18e1ed6a19f3c6ed8" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.233466 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.287016 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.304826 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.313796 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.315184 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.322835 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.333570 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.484062 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97d27\" (UniqueName: \"kubernetes.io/projected/d6e36e84-7af6-4a0b-98c8-dc100ece345b-kube-api-access-97d27\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.484129 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e36e84-7af6-4a0b-98c8-dc100ece345b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.484152 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e36e84-7af6-4a0b-98c8-dc100ece345b-config-data\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.552787 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Nov 25 10:08:57 crc kubenswrapper[4926]: W1125 10:08:57.557109 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde6e1390_0172_4a3f_b6cb_5abb2a5633b2.slice/crio-5673ca9435678568d642b33568763043b428f04bcec58fd837d55ea3604446a7 WatchSource:0}: Error finding container 5673ca9435678568d642b33568763043b428f04bcec58fd837d55ea3604446a7: Status 404 returned error can't find the container with id 5673ca9435678568d642b33568763043b428f04bcec58fd837d55ea3604446a7 Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.586499 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97d27\" (UniqueName: \"kubernetes.io/projected/d6e36e84-7af6-4a0b-98c8-dc100ece345b-kube-api-access-97d27\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.586579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e36e84-7af6-4a0b-98c8-dc100ece345b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.586602 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e36e84-7af6-4a0b-98c8-dc100ece345b-config-data\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.592565 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d6e36e84-7af6-4a0b-98c8-dc100ece345b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.593515 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6e36e84-7af6-4a0b-98c8-dc100ece345b-config-data\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.606626 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97d27\" (UniqueName: \"kubernetes.io/projected/d6e36e84-7af6-4a0b-98c8-dc100ece345b-kube-api-access-97d27\") pod \"nova-scheduler-0\" (UID: \"d6e36e84-7af6-4a0b-98c8-dc100ece345b\") " pod="openstack/nova-scheduler-0" Nov 25 10:08:57 crc kubenswrapper[4926]: I1125 10:08:57.647412 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.024086 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3dc1ea4b-a014-4954-bbfc-becf710f80ee" path="/var/lib/kubelet/pods/3dc1ea4b-a014-4954-bbfc-becf710f80ee/volumes" Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.025378 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="406430ac-1468-495b-a177-52e52f5a0405" path="/var/lib/kubelet/pods/406430ac-1468-495b-a177-52e52f5a0405/volumes" Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.101833 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Nov 25 10:08:58 crc kubenswrapper[4926]: W1125 10:08:58.103473 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6e36e84_7af6_4a0b_98c8_dc100ece345b.slice/crio-07d1a714a761504def699bd338da8c16a02ba2c5daaf54fcda069e65fd2610c6 WatchSource:0}: Error finding container 07d1a714a761504def699bd338da8c16a02ba2c5daaf54fcda069e65fd2610c6: Status 404 returned error can't find the container with id 07d1a714a761504def699bd338da8c16a02ba2c5daaf54fcda069e65fd2610c6 Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.248425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de6e1390-0172-4a3f-b6cb-5abb2a5633b2","Type":"ContainerStarted","Data":"1f10531ad2dda1101126b3b202e6baeb984a56dbd52309974eaf080ca66c1fd6"} Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.248768 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de6e1390-0172-4a3f-b6cb-5abb2a5633b2","Type":"ContainerStarted","Data":"c414078bc534d5a10c4d824c50290313e0ec37b0ffdb6d9ee3fde2cdb535baab"} Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.248787 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de6e1390-0172-4a3f-b6cb-5abb2a5633b2","Type":"ContainerStarted","Data":"5673ca9435678568d642b33568763043b428f04bcec58fd837d55ea3604446a7"} Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.251646 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6e36e84-7af6-4a0b-98c8-dc100ece345b","Type":"ContainerStarted","Data":"07d1a714a761504def699bd338da8c16a02ba2c5daaf54fcda069e65fd2610c6"} Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.278447 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.278420893 podStartE2EDuration="2.278420893s" podCreationTimestamp="2025-11-25 10:08:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:58.269920381 +0000 UTC m=+1234.953876818" watchObservedRunningTime="2025-11-25 10:08:58.278420893 +0000 UTC m=+1234.962377320" Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.817075 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": read tcp 10.217.0.2:56106->10.217.0.176:8775: read: connection reset by peer" Nov 25 10:08:58 crc kubenswrapper[4926]: I1125 10:08:58.817461 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": read tcp 10.217.0.2:56110->10.217.0.176:8775: read: connection reset by peer" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.255985 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.260833 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"d6e36e84-7af6-4a0b-98c8-dc100ece345b","Type":"ContainerStarted","Data":"e3da294f2f6347cd199d6e600dbf7a361e13176a5f7b9c28d23ab1154558f143"} Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.262806 4926 generic.go:334] "Generic (PLEG): container finished" podID="1760dcfa-c766-4bf2-b054-223863000927" containerID="3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711" exitCode=0 Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.262851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1760dcfa-c766-4bf2-b054-223863000927","Type":"ContainerDied","Data":"3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711"} Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.262893 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.262915 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"1760dcfa-c766-4bf2-b054-223863000927","Type":"ContainerDied","Data":"afb2443e09c16ae6bb12851dcb3f18ea7ff51df1e5feb31d079e602785d29595"} Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.262959 4926 scope.go:117] "RemoveContainer" containerID="3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.312814 4926 scope.go:117] "RemoveContainer" containerID="61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.318284 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.318265921 podStartE2EDuration="2.318265921s" podCreationTimestamp="2025-11-25 10:08:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:08:59.297630767 +0000 UTC m=+1235.981587204" watchObservedRunningTime="2025-11-25 10:08:59.318265921 +0000 UTC m=+1236.002222348" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.353664 4926 scope.go:117] "RemoveContainer" containerID="3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711" Nov 25 10:08:59 crc kubenswrapper[4926]: E1125 10:08:59.354091 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711\": container with ID starting with 3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711 not found: ID does not exist" containerID="3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.354129 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711"} err="failed to get container status \"3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711\": rpc error: code = NotFound desc = could not find container \"3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711\": container with ID starting with 3c8426d371aafb0809745e8c36576542237b7fa8d134a1e508897a055890b711 not found: ID does not exist" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.354154 4926 scope.go:117] "RemoveContainer" containerID="61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9" Nov 25 10:08:59 crc kubenswrapper[4926]: E1125 10:08:59.355028 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9\": container with ID starting with 61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9 not found: ID does not exist" containerID="61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.355056 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9"} err="failed to get container status \"61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9\": rpc error: code = NotFound desc = could not find container \"61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9\": container with ID starting with 61e3ad808c6834230f0535d74b7a645510080d574449f49664c310cab86e67e9 not found: ID does not exist" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.417412 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-combined-ca-bundle\") pod \"1760dcfa-c766-4bf2-b054-223863000927\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.417456 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-config-data\") pod \"1760dcfa-c766-4bf2-b054-223863000927\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.417499 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1760dcfa-c766-4bf2-b054-223863000927-logs\") pod \"1760dcfa-c766-4bf2-b054-223863000927\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.417604 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvd2f\" (UniqueName: \"kubernetes.io/projected/1760dcfa-c766-4bf2-b054-223863000927-kube-api-access-gvd2f\") pod \"1760dcfa-c766-4bf2-b054-223863000927\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.417642 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-nova-metadata-tls-certs\") pod \"1760dcfa-c766-4bf2-b054-223863000927\" (UID: \"1760dcfa-c766-4bf2-b054-223863000927\") " Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.418524 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1760dcfa-c766-4bf2-b054-223863000927-logs" (OuterVolumeSpecName: "logs") pod "1760dcfa-c766-4bf2-b054-223863000927" (UID: "1760dcfa-c766-4bf2-b054-223863000927"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.423313 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1760dcfa-c766-4bf2-b054-223863000927-kube-api-access-gvd2f" (OuterVolumeSpecName: "kube-api-access-gvd2f") pod "1760dcfa-c766-4bf2-b054-223863000927" (UID: "1760dcfa-c766-4bf2-b054-223863000927"). InnerVolumeSpecName "kube-api-access-gvd2f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.449358 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1760dcfa-c766-4bf2-b054-223863000927" (UID: "1760dcfa-c766-4bf2-b054-223863000927"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.465097 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-config-data" (OuterVolumeSpecName: "config-data") pod "1760dcfa-c766-4bf2-b054-223863000927" (UID: "1760dcfa-c766-4bf2-b054-223863000927"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.469326 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "1760dcfa-c766-4bf2-b054-223863000927" (UID: "1760dcfa-c766-4bf2-b054-223863000927"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.519505 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvd2f\" (UniqueName: \"kubernetes.io/projected/1760dcfa-c766-4bf2-b054-223863000927-kube-api-access-gvd2f\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.519541 4926 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.519551 4926 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.519561 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1760dcfa-c766-4bf2-b054-223863000927-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.519571 4926 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1760dcfa-c766-4bf2-b054-223863000927-logs\") on node \"crc\" DevicePath \"\"" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.614108 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.630774 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.644798 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:59 crc kubenswrapper[4926]: E1125 10:08:59.645269 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-log" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.645289 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-log" Nov 25 10:08:59 crc kubenswrapper[4926]: E1125 10:08:59.645307 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-metadata" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.645315 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-metadata" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.645562 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-metadata" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.645588 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1760dcfa-c766-4bf2-b054-223863000927" containerName="nova-metadata-log" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.646796 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.648375 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.649115 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.657047 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.722445 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.722497 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1abc2c8-9361-47b7-88f1-818915a5800e-logs\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.722528 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-config-data\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.722623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvkqb\" (UniqueName: \"kubernetes.io/projected/e1abc2c8-9361-47b7-88f1-818915a5800e-kube-api-access-xvkqb\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.722645 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.823755 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvkqb\" (UniqueName: \"kubernetes.io/projected/e1abc2c8-9361-47b7-88f1-818915a5800e-kube-api-access-xvkqb\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.823802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.823856 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.823881 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1abc2c8-9361-47b7-88f1-818915a5800e-logs\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.823908 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-config-data\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.824519 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1abc2c8-9361-47b7-88f1-818915a5800e-logs\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.827832 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.828014 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.835067 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1abc2c8-9361-47b7-88f1-818915a5800e-config-data\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.839374 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvkqb\" (UniqueName: \"kubernetes.io/projected/e1abc2c8-9361-47b7-88f1-818915a5800e-kube-api-access-xvkqb\") pod \"nova-metadata-0\" (UID: \"e1abc2c8-9361-47b7-88f1-818915a5800e\") " pod="openstack/nova-metadata-0" Nov 25 10:08:59 crc kubenswrapper[4926]: I1125 10:08:59.973046 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Nov 25 10:09:00 crc kubenswrapper[4926]: I1125 10:09:00.035060 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1760dcfa-c766-4bf2-b054-223863000927" path="/var/lib/kubelet/pods/1760dcfa-c766-4bf2-b054-223863000927/volumes" Nov 25 10:09:00 crc kubenswrapper[4926]: W1125 10:09:00.452237 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1abc2c8_9361_47b7_88f1_818915a5800e.slice/crio-aa49b886f105985b0d847e044a916c2855b1b3e828c287c92ced1cdf486f406a WatchSource:0}: Error finding container aa49b886f105985b0d847e044a916c2855b1b3e828c287c92ced1cdf486f406a: Status 404 returned error can't find the container with id aa49b886f105985b0d847e044a916c2855b1b3e828c287c92ced1cdf486f406a Nov 25 10:09:00 crc kubenswrapper[4926]: I1125 10:09:00.457768 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Nov 25 10:09:01 crc kubenswrapper[4926]: I1125 10:09:01.286749 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1abc2c8-9361-47b7-88f1-818915a5800e","Type":"ContainerStarted","Data":"44161b4937f2431b6ffd3a07d011733fef6514e3543fb405f39703c341ea8665"} Nov 25 10:09:01 crc kubenswrapper[4926]: I1125 10:09:01.287165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1abc2c8-9361-47b7-88f1-818915a5800e","Type":"ContainerStarted","Data":"ffa9244e78c69f1f2d512f373bcefb4a6acfcbb310c08c7b1e6ab19868381b0e"} Nov 25 10:09:01 crc kubenswrapper[4926]: I1125 10:09:01.287182 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e1abc2c8-9361-47b7-88f1-818915a5800e","Type":"ContainerStarted","Data":"aa49b886f105985b0d847e044a916c2855b1b3e828c287c92ced1cdf486f406a"} Nov 25 10:09:01 crc kubenswrapper[4926]: I1125 10:09:01.309225 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.309200939 podStartE2EDuration="2.309200939s" podCreationTimestamp="2025-11-25 10:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:09:01.307473166 +0000 UTC m=+1237.991429633" watchObservedRunningTime="2025-11-25 10:09:01.309200939 +0000 UTC m=+1237.993157376" Nov 25 10:09:02 crc kubenswrapper[4926]: I1125 10:09:02.648040 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Nov 25 10:09:04 crc kubenswrapper[4926]: I1125 10:09:04.973643 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:09:04 crc kubenswrapper[4926]: I1125 10:09:04.974086 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Nov 25 10:09:06 crc kubenswrapper[4926]: I1125 10:09:06.931031 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:09:06 crc kubenswrapper[4926]: I1125 10:09:06.931360 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Nov 25 10:09:07 crc kubenswrapper[4926]: I1125 10:09:07.647608 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Nov 25 10:09:07 crc kubenswrapper[4926]: I1125 10:09:07.683405 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Nov 25 10:09:07 crc kubenswrapper[4926]: I1125 10:09:07.947165 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de6e1390-0172-4a3f-b6cb-5abb2a5633b2" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.184:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:09:07 crc kubenswrapper[4926]: I1125 10:09:07.947520 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de6e1390-0172-4a3f-b6cb-5abb2a5633b2" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.184:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:09:08 crc kubenswrapper[4926]: I1125 10:09:08.410949 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Nov 25 10:09:09 crc kubenswrapper[4926]: I1125 10:09:09.973733 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:09:09 crc kubenswrapper[4926]: I1125 10:09:09.973822 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Nov 25 10:09:10 crc kubenswrapper[4926]: I1125 10:09:10.990134 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e1abc2c8-9361-47b7-88f1-818915a5800e" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 25 10:09:10 crc kubenswrapper[4926]: I1125 10:09:10.990140 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e1abc2c8-9361-47b7-88f1-818915a5800e" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.186:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Nov 25 10:09:16 crc kubenswrapper[4926]: I1125 10:09:16.944397 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:09:16 crc kubenswrapper[4926]: I1125 10:09:16.944989 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Nov 25 10:09:16 crc kubenswrapper[4926]: I1125 10:09:16.945651 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:09:16 crc kubenswrapper[4926]: I1125 10:09:16.945707 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Nov 25 10:09:16 crc kubenswrapper[4926]: I1125 10:09:16.958250 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:09:16 crc kubenswrapper[4926]: I1125 10:09:16.958342 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Nov 25 10:09:19 crc kubenswrapper[4926]: I1125 10:09:19.980394 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:09:19 crc kubenswrapper[4926]: I1125 10:09:19.985457 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Nov 25 10:09:19 crc kubenswrapper[4926]: I1125 10:09:19.988914 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:09:20 crc kubenswrapper[4926]: I1125 10:09:20.480622 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Nov 25 10:09:20 crc kubenswrapper[4926]: I1125 10:09:20.587260 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Nov 25 10:09:29 crc kubenswrapper[4926]: I1125 10:09:29.696244 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:09:31 crc kubenswrapper[4926]: I1125 10:09:31.323815 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:09:33 crc kubenswrapper[4926]: I1125 10:09:33.494795 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="rabbitmq" containerID="cri-o://caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380" gracePeriod=604797 Nov 25 10:09:34 crc kubenswrapper[4926]: I1125 10:09:34.633902 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Nov 25 10:09:35 crc kubenswrapper[4926]: I1125 10:09:35.447496 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerName="rabbitmq" containerID="cri-o://4a21748c09e0a6d16c38b18f9e819a289276f74dcd6dc74be0b5b0bac57419a5" gracePeriod=604796 Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.059678 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202453 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64dbdcf7-3a3c-48c3-a68f-22644f250edd-erlang-cookie-secret\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202509 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64dbdcf7-3a3c-48c3-a68f-22644f250edd-pod-info\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202611 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202744 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-server-conf\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202765 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-erlang-cookie\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202801 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-config-data\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202826 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-plugins-conf\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202852 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-plugins\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202883 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-tls\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202902 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grgn6\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-kube-api-access-grgn6\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.202919 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-confd\") pod \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\" (UID: \"64dbdcf7-3a3c-48c3-a68f-22644f250edd\") " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.204109 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.206298 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.206608 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.221668 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64dbdcf7-3a3c-48c3-a68f-22644f250edd-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.222153 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-kube-api-access-grgn6" (OuterVolumeSpecName: "kube-api-access-grgn6") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "kube-api-access-grgn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.223605 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/64dbdcf7-3a3c-48c3-a68f-22644f250edd-pod-info" (OuterVolumeSpecName: "pod-info") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.226455 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.231113 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.235533 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-config-data" (OuterVolumeSpecName: "config-data") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.281489 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-server-conf" (OuterVolumeSpecName: "server-conf") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304710 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304739 4926 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304751 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304761 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304769 4926 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/64dbdcf7-3a3c-48c3-a68f-22644f250edd-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304777 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304785 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304794 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grgn6\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-kube-api-access-grgn6\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304804 4926 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/64dbdcf7-3a3c-48c3-a68f-22644f250edd-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.304812 4926 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/64dbdcf7-3a3c-48c3-a68f-22644f250edd-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.329396 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.360032 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "64dbdcf7-3a3c-48c3-a68f-22644f250edd" (UID: "64dbdcf7-3a3c-48c3-a68f-22644f250edd"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.406755 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/64dbdcf7-3a3c-48c3-a68f-22644f250edd-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.406798 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.675928 4926 generic.go:334] "Generic (PLEG): container finished" podID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerID="caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380" exitCode=0 Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.675988 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"64dbdcf7-3a3c-48c3-a68f-22644f250edd","Type":"ContainerDied","Data":"caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380"} Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.676015 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"64dbdcf7-3a3c-48c3-a68f-22644f250edd","Type":"ContainerDied","Data":"e23c16e3d6fc2c9d531dc78028ed4c4b1663c115d3156f4b0597313ca1684e6e"} Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.676021 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.676033 4926 scope.go:117] "RemoveContainer" containerID="caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.707912 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.708413 4926 scope.go:117] "RemoveContainer" containerID="07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.720288 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.743106 4926 scope.go:117] "RemoveContainer" containerID="caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380" Nov 25 10:09:40 crc kubenswrapper[4926]: E1125 10:09:40.743999 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380\": container with ID starting with caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380 not found: ID does not exist" containerID="caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.744038 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380"} err="failed to get container status \"caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380\": rpc error: code = NotFound desc = could not find container \"caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380\": container with ID starting with caabcbb571a6f909167694e361087f1744e3ffe2c4309557ef45c7eb4558d380 not found: ID does not exist" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.744060 4926 scope.go:117] "RemoveContainer" containerID="07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b" Nov 25 10:09:40 crc kubenswrapper[4926]: E1125 10:09:40.749051 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b\": container with ID starting with 07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b not found: ID does not exist" containerID="07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.749087 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b"} err="failed to get container status \"07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b\": rpc error: code = NotFound desc = could not find container \"07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b\": container with ID starting with 07897ec17646c871ddac52c148635b840ab9b340c2c5d662844a564c1eb7d67b not found: ID does not exist" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.750760 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:09:40 crc kubenswrapper[4926]: E1125 10:09:40.751550 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="rabbitmq" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.751572 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="rabbitmq" Nov 25 10:09:40 crc kubenswrapper[4926]: E1125 10:09:40.751609 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="setup-container" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.751618 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="setup-container" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.751826 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" containerName="rabbitmq" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.753428 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.764713 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.816890 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.817400 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.817687 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.817709 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ddhxn" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.819987 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.820008 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.820082 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.916121 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.916425 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.916555 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.916670 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.916820 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-config-data\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.916952 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.917049 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.917165 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.917299 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7pnn\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-kube-api-access-t7pnn\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.917354 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:40 crc kubenswrapper[4926]: I1125 10:09:40.917381 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019479 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019531 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019579 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019603 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7pnn\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-kube-api-access-t7pnn\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019645 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019715 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019746 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019787 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019822 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.019875 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-config-data\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.020423 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.020765 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.021068 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-config-data\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.021334 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-server-conf\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.021349 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.021581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.025528 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.027575 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.027752 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.037208 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-pod-info\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.040654 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7pnn\" (UniqueName: \"kubernetes.io/projected/4bccf53f-5bb6-4e30-876e-e3d6be02f9fe-kube-api-access-t7pnn\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.056030 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe\") " pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.136860 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.571549 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Nov 25 10:09:41 crc kubenswrapper[4926]: W1125 10:09:41.572976 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4bccf53f_5bb6_4e30_876e_e3d6be02f9fe.slice/crio-f6128b9fa6608b7df66a2e51517725c36d496ee45f33ee3efd2a30bb319381dc WatchSource:0}: Error finding container f6128b9fa6608b7df66a2e51517725c36d496ee45f33ee3efd2a30bb319381dc: Status 404 returned error can't find the container with id f6128b9fa6608b7df66a2e51517725c36d496ee45f33ee3efd2a30bb319381dc Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.687045 4926 generic.go:334] "Generic (PLEG): container finished" podID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerID="4a21748c09e0a6d16c38b18f9e819a289276f74dcd6dc74be0b5b0bac57419a5" exitCode=0 Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.687103 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78f39106-b4c8-41d0-b852-a12e64f5863b","Type":"ContainerDied","Data":"4a21748c09e0a6d16c38b18f9e819a289276f74dcd6dc74be0b5b0bac57419a5"} Nov 25 10:09:41 crc kubenswrapper[4926]: I1125 10:09:41.692119 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe","Type":"ContainerStarted","Data":"f6128b9fa6608b7df66a2e51517725c36d496ee45f33ee3efd2a30bb319381dc"} Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.038566 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.042112 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64dbdcf7-3a3c-48c3-a68f-22644f250edd" path="/var/lib/kubelet/pods/64dbdcf7-3a3c-48c3-a68f-22644f250edd/volumes" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145093 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-plugins-conf\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145161 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-plugins\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145210 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qg2n\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-kube-api-access-6qg2n\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145260 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-tls\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145294 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145436 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78f39106-b4c8-41d0-b852-a12e64f5863b-erlang-cookie-secret\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145471 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-config-data\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145512 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-erlang-cookie\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145579 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-server-conf\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145620 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-confd\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.145670 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78f39106-b4c8-41d0-b852-a12e64f5863b-pod-info\") pod \"78f39106-b4c8-41d0-b852-a12e64f5863b\" (UID: \"78f39106-b4c8-41d0-b852-a12e64f5863b\") " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.147080 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.147105 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.149268 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.154131 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "persistence") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.154150 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-kube-api-access-6qg2n" (OuterVolumeSpecName: "kube-api-access-6qg2n") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "kube-api-access-6qg2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.154146 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78f39106-b4c8-41d0-b852-a12e64f5863b-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.159228 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/78f39106-b4c8-41d0-b852-a12e64f5863b-pod-info" (OuterVolumeSpecName: "pod-info") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.177132 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.180060 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-config-data" (OuterVolumeSpecName: "config-data") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.205856 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-server-conf" (OuterVolumeSpecName: "server-conf") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248204 4926 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/78f39106-b4c8-41d0-b852-a12e64f5863b-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248245 4926 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-config-data\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248260 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248275 4926 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-server-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248286 4926 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/78f39106-b4c8-41d0-b852-a12e64f5863b-pod-info\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248296 4926 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/78f39106-b4c8-41d0-b852-a12e64f5863b-plugins-conf\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248307 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248317 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qg2n\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-kube-api-access-6qg2n\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248330 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.248357 4926 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.272633 4926 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.349769 4926 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.377118 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "78f39106-b4c8-41d0-b852-a12e64f5863b" (UID: "78f39106-b4c8-41d0-b852-a12e64f5863b"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.451635 4926 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/78f39106-b4c8-41d0-b852-a12e64f5863b-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.702779 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"78f39106-b4c8-41d0-b852-a12e64f5863b","Type":"ContainerDied","Data":"6bd78357f61cc706e1109a39b225e8e3b5c0bf417d721046b11af1a0957312b6"} Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.703442 4926 scope.go:117] "RemoveContainer" containerID="4a21748c09e0a6d16c38b18f9e819a289276f74dcd6dc74be0b5b0bac57419a5" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.702822 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.734924 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.735738 4926 scope.go:117] "RemoveContainer" containerID="838134005ec8486202b3ca3e5e28bed22a4b79db2906d9166c0e208a2fcef3c5" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.742975 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.781713 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:09:42 crc kubenswrapper[4926]: E1125 10:09:42.782235 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerName="setup-container" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.782248 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerName="setup-container" Nov 25 10:09:42 crc kubenswrapper[4926]: E1125 10:09:42.782274 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerName="rabbitmq" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.782280 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerName="rabbitmq" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.783494 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" containerName="rabbitmq" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.784858 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.796333 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.796544 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.796711 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.796839 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.797022 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.797239 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.799160 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4plth" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.804559 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967053 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967101 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967137 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt5kn\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-kube-api-access-qt5kn\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967193 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967216 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967238 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967257 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967299 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967317 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:42 crc kubenswrapper[4926]: I1125 10:09:42.967417 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070182 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070236 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070275 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt5kn\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-kube-api-access-qt5kn\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070311 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070345 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070374 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070403 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070462 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070488 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070523 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.070566 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.071772 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.072072 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.072407 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.072601 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.072656 4926 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.073128 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.076672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.077251 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.078235 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.078367 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.092700 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt5kn\" (UniqueName: \"kubernetes.io/projected/30c23e52-ddbc-4e38-85f6-6a7b32e67d12-kube-api-access-qt5kn\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.116695 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"30c23e52-ddbc-4e38-85f6-6a7b32e67d12\") " pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.252187 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.715368 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe","Type":"ContainerStarted","Data":"81d6b1e11aa9471682106db607c8d87a1c58d8fe297b62eb91424bba6b11146a"} Nov 25 10:09:43 crc kubenswrapper[4926]: W1125 10:09:43.756646 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod30c23e52_ddbc_4e38_85f6_6a7b32e67d12.slice/crio-c33d5ac84ed5a017fc642222b6b0045969dbd20fec290236b5eb753147ff9362 WatchSource:0}: Error finding container c33d5ac84ed5a017fc642222b6b0045969dbd20fec290236b5eb753147ff9362: Status 404 returned error can't find the container with id c33d5ac84ed5a017fc642222b6b0045969dbd20fec290236b5eb753147ff9362 Nov 25 10:09:43 crc kubenswrapper[4926]: I1125 10:09:43.760962 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.018255 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.018623 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.023909 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78f39106-b4c8-41d0-b852-a12e64f5863b" path="/var/lib/kubelet/pods/78f39106-b4c8-41d0-b852-a12e64f5863b/volumes" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.461581 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-njv6x"] Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.464911 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.472851 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.478045 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-njv6x"] Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.597600 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-dns-svc\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.597661 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggf49\" (UniqueName: \"kubernetes.io/projected/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-kube-api-access-ggf49\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.597698 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.597783 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.597839 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.597907 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-config\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.698914 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-config\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.699235 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-dns-svc\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.699278 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggf49\" (UniqueName: \"kubernetes.io/projected/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-kube-api-access-ggf49\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.699307 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.699388 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.699437 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.700293 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-openstack-edpm-ipam\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.700826 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-sb\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.700856 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-nb\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.700926 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-dns-svc\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.701078 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-config\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.723667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30c23e52-ddbc-4e38-85f6-6a7b32e67d12","Type":"ContainerStarted","Data":"c33d5ac84ed5a017fc642222b6b0045969dbd20fec290236b5eb753147ff9362"} Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.753263 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggf49\" (UniqueName: \"kubernetes.io/projected/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-kube-api-access-ggf49\") pod \"dnsmasq-dns-578b8d767c-njv6x\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:44 crc kubenswrapper[4926]: I1125 10:09:44.786558 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:45 crc kubenswrapper[4926]: I1125 10:09:45.232309 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-njv6x"] Nov 25 10:09:45 crc kubenswrapper[4926]: I1125 10:09:45.733798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30c23e52-ddbc-4e38-85f6-6a7b32e67d12","Type":"ContainerStarted","Data":"68da5a6fdad3618fc88f1ce53faa867a4b1d6f5346e8caaa5db2bdf41af6fb4e"} Nov 25 10:09:45 crc kubenswrapper[4926]: I1125 10:09:45.737010 4926 generic.go:334] "Generic (PLEG): container finished" podID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerID="e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209" exitCode=0 Nov 25 10:09:45 crc kubenswrapper[4926]: I1125 10:09:45.737054 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" event={"ID":"e344cd1d-f76f-4577-a9ea-5f886fcc86aa","Type":"ContainerDied","Data":"e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209"} Nov 25 10:09:45 crc kubenswrapper[4926]: I1125 10:09:45.737082 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" event={"ID":"e344cd1d-f76f-4577-a9ea-5f886fcc86aa","Type":"ContainerStarted","Data":"76c584a1df3127662e3215c2c1cbd162b11cb4304a3a2ed5512a4349170072c8"} Nov 25 10:09:46 crc kubenswrapper[4926]: I1125 10:09:46.765700 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" event={"ID":"e344cd1d-f76f-4577-a9ea-5f886fcc86aa","Type":"ContainerStarted","Data":"125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535"} Nov 25 10:09:46 crc kubenswrapper[4926]: I1125 10:09:46.765991 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:46 crc kubenswrapper[4926]: I1125 10:09:46.792618 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" podStartSLOduration=2.79260341 podStartE2EDuration="2.79260341s" podCreationTimestamp="2025-11-25 10:09:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:09:46.783655887 +0000 UTC m=+1283.467612314" watchObservedRunningTime="2025-11-25 10:09:46.79260341 +0000 UTC m=+1283.476559837" Nov 25 10:09:54 crc kubenswrapper[4926]: I1125 10:09:54.789095 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:09:54 crc kubenswrapper[4926]: I1125 10:09:54.870772 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-sbmk8"] Nov 25 10:09:54 crc kubenswrapper[4926]: I1125 10:09:54.871074 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerName="dnsmasq-dns" containerID="cri-o://7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401" gracePeriod=10 Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.040198 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-fbc59fbb7-nf2mz"] Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.042108 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.062080 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fbc59fbb7-nf2mz"] Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.095099 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-ovsdbserver-sb\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.095138 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-dns-svc\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.095159 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-config\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.095186 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-openstack-edpm-ipam\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.095301 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rp7vm\" (UniqueName: \"kubernetes.io/projected/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-kube-api-access-rp7vm\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.095351 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-ovsdbserver-nb\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.197541 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rp7vm\" (UniqueName: \"kubernetes.io/projected/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-kube-api-access-rp7vm\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.197608 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-ovsdbserver-nb\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.197657 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-ovsdbserver-sb\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.197674 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-dns-svc\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.197689 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-config\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.197709 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-openstack-edpm-ipam\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.198467 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-openstack-edpm-ipam\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.199672 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-ovsdbserver-nb\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.200086 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-dns-svc\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.200285 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-ovsdbserver-sb\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.201019 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-config\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.229307 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rp7vm\" (UniqueName: \"kubernetes.io/projected/9252a64e-adab-4db0-aa4e-6bcbb0b9458b-kube-api-access-rp7vm\") pod \"dnsmasq-dns-fbc59fbb7-nf2mz\" (UID: \"9252a64e-adab-4db0-aa4e-6bcbb0b9458b\") " pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.384230 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.454218 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.504559 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-sb\") pod \"1a66f16d-89fd-4885-bc00-87f73e6705e8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.504699 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bd5xf\" (UniqueName: \"kubernetes.io/projected/1a66f16d-89fd-4885-bc00-87f73e6705e8-kube-api-access-bd5xf\") pod \"1a66f16d-89fd-4885-bc00-87f73e6705e8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.504731 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-config\") pod \"1a66f16d-89fd-4885-bc00-87f73e6705e8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.504769 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-nb\") pod \"1a66f16d-89fd-4885-bc00-87f73e6705e8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.504802 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-dns-svc\") pod \"1a66f16d-89fd-4885-bc00-87f73e6705e8\" (UID: \"1a66f16d-89fd-4885-bc00-87f73e6705e8\") " Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.513688 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a66f16d-89fd-4885-bc00-87f73e6705e8-kube-api-access-bd5xf" (OuterVolumeSpecName: "kube-api-access-bd5xf") pod "1a66f16d-89fd-4885-bc00-87f73e6705e8" (UID: "1a66f16d-89fd-4885-bc00-87f73e6705e8"). InnerVolumeSpecName "kube-api-access-bd5xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.570027 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1a66f16d-89fd-4885-bc00-87f73e6705e8" (UID: "1a66f16d-89fd-4885-bc00-87f73e6705e8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.570960 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1a66f16d-89fd-4885-bc00-87f73e6705e8" (UID: "1a66f16d-89fd-4885-bc00-87f73e6705e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.583575 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1a66f16d-89fd-4885-bc00-87f73e6705e8" (UID: "1a66f16d-89fd-4885-bc00-87f73e6705e8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.588674 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-config" (OuterVolumeSpecName: "config") pod "1a66f16d-89fd-4885-bc00-87f73e6705e8" (UID: "1a66f16d-89fd-4885-bc00-87f73e6705e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.607549 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.607595 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bd5xf\" (UniqueName: \"kubernetes.io/projected/1a66f16d-89fd-4885-bc00-87f73e6705e8-kube-api-access-bd5xf\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.607617 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.607634 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.607645 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1a66f16d-89fd-4885-bc00-87f73e6705e8-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.845465 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-fbc59fbb7-nf2mz"] Nov 25 10:09:55 crc kubenswrapper[4926]: W1125 10:09:55.850869 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9252a64e_adab_4db0_aa4e_6bcbb0b9458b.slice/crio-d17827b17c2400d042a9626c329be487abae13ffe854e7cc812203bcdcbbe8a9 WatchSource:0}: Error finding container d17827b17c2400d042a9626c329be487abae13ffe854e7cc812203bcdcbbe8a9: Status 404 returned error can't find the container with id d17827b17c2400d042a9626c329be487abae13ffe854e7cc812203bcdcbbe8a9 Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.855197 4926 generic.go:334] "Generic (PLEG): container finished" podID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerID="7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401" exitCode=0 Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.855235 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" event={"ID":"1a66f16d-89fd-4885-bc00-87f73e6705e8","Type":"ContainerDied","Data":"7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401"} Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.855261 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" event={"ID":"1a66f16d-89fd-4885-bc00-87f73e6705e8","Type":"ContainerDied","Data":"84647ade3f46e4e1ca2fd993772b85895e1f062b43a0a112473e4e07d41044c4"} Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.855275 4926 scope.go:117] "RemoveContainer" containerID="7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.855380 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68d4b6d797-sbmk8" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.913861 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-sbmk8"] Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.930093 4926 scope.go:117] "RemoveContainer" containerID="dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.931829 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68d4b6d797-sbmk8"] Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.969974 4926 scope.go:117] "RemoveContainer" containerID="7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401" Nov 25 10:09:55 crc kubenswrapper[4926]: E1125 10:09:55.973299 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401\": container with ID starting with 7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401 not found: ID does not exist" containerID="7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.973366 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401"} err="failed to get container status \"7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401\": rpc error: code = NotFound desc = could not find container \"7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401\": container with ID starting with 7acc1a082eb6dd75bbad0f191339dc57c0a91cf4c0c4ed6336da24ac06499401 not found: ID does not exist" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.973398 4926 scope.go:117] "RemoveContainer" containerID="dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b" Nov 25 10:09:55 crc kubenswrapper[4926]: E1125 10:09:55.973757 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b\": container with ID starting with dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b not found: ID does not exist" containerID="dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b" Nov 25 10:09:55 crc kubenswrapper[4926]: I1125 10:09:55.973796 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b"} err="failed to get container status \"dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b\": rpc error: code = NotFound desc = could not find container \"dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b\": container with ID starting with dbc2da6ead62bbe64783dd3cc8501e02541493936e3cfb6ab5f8e4c49d0a7b9b not found: ID does not exist" Nov 25 10:09:56 crc kubenswrapper[4926]: I1125 10:09:56.023887 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" path="/var/lib/kubelet/pods/1a66f16d-89fd-4885-bc00-87f73e6705e8/volumes" Nov 25 10:09:56 crc kubenswrapper[4926]: I1125 10:09:56.866256 4926 generic.go:334] "Generic (PLEG): container finished" podID="9252a64e-adab-4db0-aa4e-6bcbb0b9458b" containerID="f93c35a0e5e22ca15096b7ce3a40d50dc36ded7c18ad62d9b8c1c829bf6d7ed8" exitCode=0 Nov 25 10:09:56 crc kubenswrapper[4926]: I1125 10:09:56.866365 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" event={"ID":"9252a64e-adab-4db0-aa4e-6bcbb0b9458b","Type":"ContainerDied","Data":"f93c35a0e5e22ca15096b7ce3a40d50dc36ded7c18ad62d9b8c1c829bf6d7ed8"} Nov 25 10:09:56 crc kubenswrapper[4926]: I1125 10:09:56.866586 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" event={"ID":"9252a64e-adab-4db0-aa4e-6bcbb0b9458b","Type":"ContainerStarted","Data":"d17827b17c2400d042a9626c329be487abae13ffe854e7cc812203bcdcbbe8a9"} Nov 25 10:09:57 crc kubenswrapper[4926]: I1125 10:09:57.880109 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" event={"ID":"9252a64e-adab-4db0-aa4e-6bcbb0b9458b","Type":"ContainerStarted","Data":"8abd9d88bbbd1a784c7c41504f3e2984cca630cf9252a38eeadf3529a4ef5dd1"} Nov 25 10:09:57 crc kubenswrapper[4926]: I1125 10:09:57.880543 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:09:57 crc kubenswrapper[4926]: I1125 10:09:57.917421 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" podStartSLOduration=2.91740023 podStartE2EDuration="2.91740023s" podCreationTimestamp="2025-11-25 10:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:09:57.906797136 +0000 UTC m=+1294.590753563" watchObservedRunningTime="2025-11-25 10:09:57.91740023 +0000 UTC m=+1294.601356677" Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.385564 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-fbc59fbb7-nf2mz" Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.438908 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-njv6x"] Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.439158 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerName="dnsmasq-dns" containerID="cri-o://125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535" gracePeriod=10 Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.890681 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.949377 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.949468 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" event={"ID":"e344cd1d-f76f-4577-a9ea-5f886fcc86aa","Type":"ContainerDied","Data":"125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535"} Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.949543 4926 scope.go:117] "RemoveContainer" containerID="125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535" Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.949276 4926 generic.go:334] "Generic (PLEG): container finished" podID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerID="125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535" exitCode=0 Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.949898 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-578b8d767c-njv6x" event={"ID":"e344cd1d-f76f-4577-a9ea-5f886fcc86aa","Type":"ContainerDied","Data":"76c584a1df3127662e3215c2c1cbd162b11cb4304a3a2ed5512a4349170072c8"} Nov 25 10:10:05 crc kubenswrapper[4926]: I1125 10:10:05.981781 4926 scope.go:117] "RemoveContainer" containerID="e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.000472 4926 scope.go:117] "RemoveContainer" containerID="125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535" Nov 25 10:10:06 crc kubenswrapper[4926]: E1125 10:10:06.000967 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535\": container with ID starting with 125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535 not found: ID does not exist" containerID="125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.001000 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535"} err="failed to get container status \"125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535\": rpc error: code = NotFound desc = could not find container \"125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535\": container with ID starting with 125491d8e79696c48d980a591ab004627fe29f3ed33e495d280067f169bab535 not found: ID does not exist" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.001034 4926 scope.go:117] "RemoveContainer" containerID="e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209" Nov 25 10:10:06 crc kubenswrapper[4926]: E1125 10:10:06.001373 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209\": container with ID starting with e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209 not found: ID does not exist" containerID="e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.001411 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209"} err="failed to get container status \"e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209\": rpc error: code = NotFound desc = could not find container \"e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209\": container with ID starting with e83c3e90297325436d7e42f130748c1e7fd74bf58191847ef39148be0eec1209 not found: ID does not exist" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.006131 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-dns-svc\") pod \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.006217 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-sb\") pod \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.006244 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-config\") pod \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.006265 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-nb\") pod \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.006288 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggf49\" (UniqueName: \"kubernetes.io/projected/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-kube-api-access-ggf49\") pod \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.006332 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-openstack-edpm-ipam\") pod \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\" (UID: \"e344cd1d-f76f-4577-a9ea-5f886fcc86aa\") " Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.012482 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-kube-api-access-ggf49" (OuterVolumeSpecName: "kube-api-access-ggf49") pod "e344cd1d-f76f-4577-a9ea-5f886fcc86aa" (UID: "e344cd1d-f76f-4577-a9ea-5f886fcc86aa"). InnerVolumeSpecName "kube-api-access-ggf49". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.067591 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "e344cd1d-f76f-4577-a9ea-5f886fcc86aa" (UID: "e344cd1d-f76f-4577-a9ea-5f886fcc86aa"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.068395 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e344cd1d-f76f-4577-a9ea-5f886fcc86aa" (UID: "e344cd1d-f76f-4577-a9ea-5f886fcc86aa"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.071432 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-config" (OuterVolumeSpecName: "config") pod "e344cd1d-f76f-4577-a9ea-5f886fcc86aa" (UID: "e344cd1d-f76f-4577-a9ea-5f886fcc86aa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.079990 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e344cd1d-f76f-4577-a9ea-5f886fcc86aa" (UID: "e344cd1d-f76f-4577-a9ea-5f886fcc86aa"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.100165 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e344cd1d-f76f-4577-a9ea-5f886fcc86aa" (UID: "e344cd1d-f76f-4577-a9ea-5f886fcc86aa"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.113979 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.114013 4926 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-config\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.114051 4926 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.114064 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggf49\" (UniqueName: \"kubernetes.io/projected/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-kube-api-access-ggf49\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.114077 4926 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.114088 4926 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e344cd1d-f76f-4577-a9ea-5f886fcc86aa-dns-svc\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.283137 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-njv6x"] Nov 25 10:10:06 crc kubenswrapper[4926]: I1125 10:10:06.290052 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-578b8d767c-njv6x"] Nov 25 10:10:08 crc kubenswrapper[4926]: I1125 10:10:08.021895 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" path="/var/lib/kubelet/pods/e344cd1d-f76f-4577-a9ea-5f886fcc86aa/volumes" Nov 25 10:10:14 crc kubenswrapper[4926]: I1125 10:10:14.020411 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:10:14 crc kubenswrapper[4926]: I1125 10:10:14.020910 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.619083 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp"] Nov 25 10:10:15 crc kubenswrapper[4926]: E1125 10:10:15.619710 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerName="init" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.619726 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerName="init" Nov 25 10:10:15 crc kubenswrapper[4926]: E1125 10:10:15.619751 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerName="init" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.619758 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerName="init" Nov 25 10:10:15 crc kubenswrapper[4926]: E1125 10:10:15.619777 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerName="dnsmasq-dns" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.619785 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerName="dnsmasq-dns" Nov 25 10:10:15 crc kubenswrapper[4926]: E1125 10:10:15.619809 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerName="dnsmasq-dns" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.619817 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerName="dnsmasq-dns" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.620135 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a66f16d-89fd-4885-bc00-87f73e6705e8" containerName="dnsmasq-dns" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.620157 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e344cd1d-f76f-4577-a9ea-5f886fcc86aa" containerName="dnsmasq-dns" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.623046 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.625804 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.626196 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.626690 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.627894 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.630797 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp"] Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.799345 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.799504 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjnzx\" (UniqueName: \"kubernetes.io/projected/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-kube-api-access-tjnzx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.799610 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.799684 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.902192 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.902307 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjnzx\" (UniqueName: \"kubernetes.io/projected/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-kube-api-access-tjnzx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.902465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.903012 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.909328 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.910986 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.911386 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.921182 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjnzx\" (UniqueName: \"kubernetes.io/projected/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-kube-api-access-tjnzx\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:15 crc kubenswrapper[4926]: I1125 10:10:15.940391 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:16 crc kubenswrapper[4926]: I1125 10:10:16.034452 4926 generic.go:334] "Generic (PLEG): container finished" podID="4bccf53f-5bb6-4e30-876e-e3d6be02f9fe" containerID="81d6b1e11aa9471682106db607c8d87a1c58d8fe297b62eb91424bba6b11146a" exitCode=0 Nov 25 10:10:16 crc kubenswrapper[4926]: I1125 10:10:16.034746 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe","Type":"ContainerDied","Data":"81d6b1e11aa9471682106db607c8d87a1c58d8fe297b62eb91424bba6b11146a"} Nov 25 10:10:16 crc kubenswrapper[4926]: I1125 10:10:16.672500 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp"] Nov 25 10:10:16 crc kubenswrapper[4926]: W1125 10:10:16.674457 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae06cbe1_66b6_4d10_b2fe_99834e3fc42c.slice/crio-120738bdeef61e00413030d26c5794038cca6b61e2b08ea8c91bf4942607b96b WatchSource:0}: Error finding container 120738bdeef61e00413030d26c5794038cca6b61e2b08ea8c91bf4942607b96b: Status 404 returned error can't find the container with id 120738bdeef61e00413030d26c5794038cca6b61e2b08ea8c91bf4942607b96b Nov 25 10:10:16 crc kubenswrapper[4926]: I1125 10:10:16.677305 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:10:17 crc kubenswrapper[4926]: I1125 10:10:17.045038 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" event={"ID":"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c","Type":"ContainerStarted","Data":"120738bdeef61e00413030d26c5794038cca6b61e2b08ea8c91bf4942607b96b"} Nov 25 10:10:17 crc kubenswrapper[4926]: I1125 10:10:17.047373 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"4bccf53f-5bb6-4e30-876e-e3d6be02f9fe","Type":"ContainerStarted","Data":"573d83868d40857a5850a489f98a7021b3116db2c9742e0bcef46e3903b3231a"} Nov 25 10:10:17 crc kubenswrapper[4926]: I1125 10:10:17.047670 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Nov 25 10:10:17 crc kubenswrapper[4926]: I1125 10:10:17.070462 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.070443129 podStartE2EDuration="37.070443129s" podCreationTimestamp="2025-11-25 10:09:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:10:17.07005279 +0000 UTC m=+1313.754009217" watchObservedRunningTime="2025-11-25 10:10:17.070443129 +0000 UTC m=+1313.754399556" Nov 25 10:10:18 crc kubenswrapper[4926]: I1125 10:10:18.057898 4926 generic.go:334] "Generic (PLEG): container finished" podID="30c23e52-ddbc-4e38-85f6-6a7b32e67d12" containerID="68da5a6fdad3618fc88f1ce53faa867a4b1d6f5346e8caaa5db2bdf41af6fb4e" exitCode=0 Nov 25 10:10:18 crc kubenswrapper[4926]: I1125 10:10:18.057980 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30c23e52-ddbc-4e38-85f6-6a7b32e67d12","Type":"ContainerDied","Data":"68da5a6fdad3618fc88f1ce53faa867a4b1d6f5346e8caaa5db2bdf41af6fb4e"} Nov 25 10:10:20 crc kubenswrapper[4926]: I1125 10:10:20.079470 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"30c23e52-ddbc-4e38-85f6-6a7b32e67d12","Type":"ContainerStarted","Data":"29916123e4d8402d511baec6b3e88f0ef0ef98f196a46bb8a8061446eff6a3b8"} Nov 25 10:10:20 crc kubenswrapper[4926]: I1125 10:10:20.080751 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:10:20 crc kubenswrapper[4926]: I1125 10:10:20.104588 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.104564229 podStartE2EDuration="38.104564229s" podCreationTimestamp="2025-11-25 10:09:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:10:20.100143389 +0000 UTC m=+1316.784099816" watchObservedRunningTime="2025-11-25 10:10:20.104564229 +0000 UTC m=+1316.788520666" Nov 25 10:10:26 crc kubenswrapper[4926]: I1125 10:10:26.529191 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:10:27 crc kubenswrapper[4926]: I1125 10:10:27.147277 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" event={"ID":"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c","Type":"ContainerStarted","Data":"a7fd105e4729900e70f1b83c866c78a261596757bc7a0c7177c83d8a4f0f0d41"} Nov 25 10:10:27 crc kubenswrapper[4926]: I1125 10:10:27.189069 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" podStartSLOduration=2.339252912 podStartE2EDuration="12.18903578s" podCreationTimestamp="2025-11-25 10:10:15 +0000 UTC" firstStartedPulling="2025-11-25 10:10:16.677105966 +0000 UTC m=+1313.361062393" lastFinishedPulling="2025-11-25 10:10:26.526888834 +0000 UTC m=+1323.210845261" observedRunningTime="2025-11-25 10:10:27.18020739 +0000 UTC m=+1323.864163867" watchObservedRunningTime="2025-11-25 10:10:27.18903578 +0000 UTC m=+1323.872992297" Nov 25 10:10:31 crc kubenswrapper[4926]: I1125 10:10:31.142271 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Nov 25 10:10:33 crc kubenswrapper[4926]: I1125 10:10:33.257221 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Nov 25 10:10:38 crc kubenswrapper[4926]: I1125 10:10:38.254475 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" containerID="a7fd105e4729900e70f1b83c866c78a261596757bc7a0c7177c83d8a4f0f0d41" exitCode=0 Nov 25 10:10:38 crc kubenswrapper[4926]: I1125 10:10:38.254573 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" event={"ID":"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c","Type":"ContainerDied","Data":"a7fd105e4729900e70f1b83c866c78a261596757bc7a0c7177c83d8a4f0f0d41"} Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.681528 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.772821 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-repo-setup-combined-ca-bundle\") pod \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.772998 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-inventory\") pod \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.773067 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-ssh-key\") pod \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.773132 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjnzx\" (UniqueName: \"kubernetes.io/projected/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-kube-api-access-tjnzx\") pod \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\" (UID: \"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c\") " Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.778347 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-kube-api-access-tjnzx" (OuterVolumeSpecName: "kube-api-access-tjnzx") pod "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" (UID: "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c"). InnerVolumeSpecName "kube-api-access-tjnzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.778371 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" (UID: "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.798886 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-inventory" (OuterVolumeSpecName: "inventory") pod "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" (UID: "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.800862 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" (UID: "ae06cbe1-66b6-4d10-b2fe-99834e3fc42c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.875092 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.875125 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.875137 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjnzx\" (UniqueName: \"kubernetes.io/projected/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-kube-api-access-tjnzx\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:39 crc kubenswrapper[4926]: I1125 10:10:39.875150 4926 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae06cbe1-66b6-4d10-b2fe-99834e3fc42c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.288611 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" event={"ID":"ae06cbe1-66b6-4d10-b2fe-99834e3fc42c","Type":"ContainerDied","Data":"120738bdeef61e00413030d26c5794038cca6b61e2b08ea8c91bf4942607b96b"} Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.288653 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="120738bdeef61e00413030d26c5794038cca6b61e2b08ea8c91bf4942607b96b" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.288705 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.374027 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l"] Nov 25 10:10:40 crc kubenswrapper[4926]: E1125 10:10:40.374418 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.374443 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.374687 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae06cbe1-66b6-4d10-b2fe-99834e3fc42c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.375414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.379861 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.379867 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.384192 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.384197 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.399599 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l"] Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.490413 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.490700 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.490808 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwc4r\" (UniqueName: \"kubernetes.io/projected/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-kube-api-access-cwc4r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.490975 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.592675 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.592771 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.592805 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.592862 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwc4r\" (UniqueName: \"kubernetes.io/projected/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-kube-api-access-cwc4r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.596917 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.605598 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.606297 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.608157 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwc4r\" (UniqueName: \"kubernetes.io/projected/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-kube-api-access-cwc4r\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:40 crc kubenswrapper[4926]: I1125 10:10:40.695436 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:10:41 crc kubenswrapper[4926]: I1125 10:10:41.218277 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l"] Nov 25 10:10:41 crc kubenswrapper[4926]: W1125 10:10:41.223111 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac1288a0_8cc2_49b5_b9d4_f0402c00aba9.slice/crio-8b8b184198a4af75d39dcdd805871d7644a80d52a2061d4d773b478b2c48a595 WatchSource:0}: Error finding container 8b8b184198a4af75d39dcdd805871d7644a80d52a2061d4d773b478b2c48a595: Status 404 returned error can't find the container with id 8b8b184198a4af75d39dcdd805871d7644a80d52a2061d4d773b478b2c48a595 Nov 25 10:10:41 crc kubenswrapper[4926]: I1125 10:10:41.299073 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" event={"ID":"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9","Type":"ContainerStarted","Data":"8b8b184198a4af75d39dcdd805871d7644a80d52a2061d4d773b478b2c48a595"} Nov 25 10:10:42 crc kubenswrapper[4926]: I1125 10:10:42.311647 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" event={"ID":"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9","Type":"ContainerStarted","Data":"4ddbb6fda74711116d29048e3312ec8c3d69fd0ae587fe2f7168188a0c46dead"} Nov 25 10:10:42 crc kubenswrapper[4926]: I1125 10:10:42.330299 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" podStartSLOduration=1.861222419 podStartE2EDuration="2.330280158s" podCreationTimestamp="2025-11-25 10:10:40 +0000 UTC" firstStartedPulling="2025-11-25 10:10:41.227976423 +0000 UTC m=+1337.911932850" lastFinishedPulling="2025-11-25 10:10:41.697034162 +0000 UTC m=+1338.380990589" observedRunningTime="2025-11-25 10:10:42.326625797 +0000 UTC m=+1339.010582224" watchObservedRunningTime="2025-11-25 10:10:42.330280158 +0000 UTC m=+1339.014236595" Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.017823 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.017879 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.023533 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.024020 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"780e5b5fae7c7161700dd19e5dd8229ab43351e7385e07fc9c5c1c4ee3eb5f66"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.024083 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://780e5b5fae7c7161700dd19e5dd8229ab43351e7385e07fc9c5c1c4ee3eb5f66" gracePeriod=600 Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.339264 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="780e5b5fae7c7161700dd19e5dd8229ab43351e7385e07fc9c5c1c4ee3eb5f66" exitCode=0 Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.339544 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"780e5b5fae7c7161700dd19e5dd8229ab43351e7385e07fc9c5c1c4ee3eb5f66"} Nov 25 10:10:44 crc kubenswrapper[4926]: I1125 10:10:44.339740 4926 scope.go:117] "RemoveContainer" containerID="bdee8ff08446a7fcbf9166a91fe5650194045a7c04583b6368b152639d328c44" Nov 25 10:10:45 crc kubenswrapper[4926]: I1125 10:10:45.351276 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625"} Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.151014 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6rcxc"] Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.154149 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.170048 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rcxc"] Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.278965 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54x86\" (UniqueName: \"kubernetes.io/projected/e5f64b63-b874-42b3-905a-b45833ff181b-kube-api-access-54x86\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.279107 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-utilities\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.279149 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-catalog-content\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.381287 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54x86\" (UniqueName: \"kubernetes.io/projected/e5f64b63-b874-42b3-905a-b45833ff181b-kube-api-access-54x86\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.381446 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-utilities\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.381502 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-catalog-content\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.382182 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-catalog-content\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.382181 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-utilities\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.408803 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54x86\" (UniqueName: \"kubernetes.io/projected/e5f64b63-b874-42b3-905a-b45833ff181b-kube-api-access-54x86\") pod \"redhat-marketplace-6rcxc\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.484688 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:47 crc kubenswrapper[4926]: I1125 10:11:47.999072 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rcxc"] Nov 25 10:11:48 crc kubenswrapper[4926]: I1125 10:11:48.940293 4926 generic.go:334] "Generic (PLEG): container finished" podID="e5f64b63-b874-42b3-905a-b45833ff181b" containerID="928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce" exitCode=0 Nov 25 10:11:48 crc kubenswrapper[4926]: I1125 10:11:48.940404 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rcxc" event={"ID":"e5f64b63-b874-42b3-905a-b45833ff181b","Type":"ContainerDied","Data":"928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce"} Nov 25 10:11:48 crc kubenswrapper[4926]: I1125 10:11:48.940850 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rcxc" event={"ID":"e5f64b63-b874-42b3-905a-b45833ff181b","Type":"ContainerStarted","Data":"1503606d6d49dcf385fb839a4d1caac5f7e6bddfb42b29eae604731844a2b09a"} Nov 25 10:11:49 crc kubenswrapper[4926]: I1125 10:11:49.952103 4926 generic.go:334] "Generic (PLEG): container finished" podID="e5f64b63-b874-42b3-905a-b45833ff181b" containerID="a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b" exitCode=0 Nov 25 10:11:49 crc kubenswrapper[4926]: I1125 10:11:49.952202 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rcxc" event={"ID":"e5f64b63-b874-42b3-905a-b45833ff181b","Type":"ContainerDied","Data":"a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b"} Nov 25 10:11:50 crc kubenswrapper[4926]: I1125 10:11:50.963759 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rcxc" event={"ID":"e5f64b63-b874-42b3-905a-b45833ff181b","Type":"ContainerStarted","Data":"f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4"} Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.508856 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6rcxc" podStartSLOduration=5.096894906 podStartE2EDuration="6.508836372s" podCreationTimestamp="2025-11-25 10:11:47 +0000 UTC" firstStartedPulling="2025-11-25 10:11:48.942740067 +0000 UTC m=+1405.626696504" lastFinishedPulling="2025-11-25 10:11:50.354681543 +0000 UTC m=+1407.038637970" observedRunningTime="2025-11-25 10:11:50.986846241 +0000 UTC m=+1407.670802708" watchObservedRunningTime="2025-11-25 10:11:53.508836372 +0000 UTC m=+1410.192792809" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.514684 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fwc4w"] Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.517414 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.525091 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fwc4w"] Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.601932 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-catalog-content\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.602056 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k4nvb\" (UniqueName: \"kubernetes.io/projected/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-kube-api-access-k4nvb\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.602124 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-utilities\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.703775 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-utilities\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.703901 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-catalog-content\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.704326 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-utilities\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.704392 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-catalog-content\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.704537 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k4nvb\" (UniqueName: \"kubernetes.io/projected/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-kube-api-access-k4nvb\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.728140 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k4nvb\" (UniqueName: \"kubernetes.io/projected/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-kube-api-access-k4nvb\") pod \"community-operators-fwc4w\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:53 crc kubenswrapper[4926]: I1125 10:11:53.887359 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:11:54 crc kubenswrapper[4926]: I1125 10:11:54.367194 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fwc4w"] Nov 25 10:11:54 crc kubenswrapper[4926]: W1125 10:11:54.369135 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae739f4b_cb8c_4f0e_9914_8736d77f25b3.slice/crio-7384c448db31b821864e409d593d3729cf57d0d3dee16608b99129d403a33610 WatchSource:0}: Error finding container 7384c448db31b821864e409d593d3729cf57d0d3dee16608b99129d403a33610: Status 404 returned error can't find the container with id 7384c448db31b821864e409d593d3729cf57d0d3dee16608b99129d403a33610 Nov 25 10:11:55 crc kubenswrapper[4926]: I1125 10:11:55.025909 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerID="3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334" exitCode=0 Nov 25 10:11:55 crc kubenswrapper[4926]: I1125 10:11:55.025982 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerDied","Data":"3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334"} Nov 25 10:11:55 crc kubenswrapper[4926]: I1125 10:11:55.026010 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerStarted","Data":"7384c448db31b821864e409d593d3729cf57d0d3dee16608b99129d403a33610"} Nov 25 10:11:56 crc kubenswrapper[4926]: I1125 10:11:56.037907 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerStarted","Data":"3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40"} Nov 25 10:11:57 crc kubenswrapper[4926]: I1125 10:11:57.047742 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerID="3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40" exitCode=0 Nov 25 10:11:57 crc kubenswrapper[4926]: I1125 10:11:57.047851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerDied","Data":"3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40"} Nov 25 10:11:57 crc kubenswrapper[4926]: I1125 10:11:57.485009 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:57 crc kubenswrapper[4926]: I1125 10:11:57.485096 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:57 crc kubenswrapper[4926]: I1125 10:11:57.545009 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:58 crc kubenswrapper[4926]: I1125 10:11:58.059646 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerStarted","Data":"d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8"} Nov 25 10:11:58 crc kubenswrapper[4926]: I1125 10:11:58.083375 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fwc4w" podStartSLOduration=2.577182518 podStartE2EDuration="5.083345785s" podCreationTimestamp="2025-11-25 10:11:53 +0000 UTC" firstStartedPulling="2025-11-25 10:11:55.027605422 +0000 UTC m=+1411.711561849" lastFinishedPulling="2025-11-25 10:11:57.533768679 +0000 UTC m=+1414.217725116" observedRunningTime="2025-11-25 10:11:58.076856724 +0000 UTC m=+1414.760813151" watchObservedRunningTime="2025-11-25 10:11:58.083345785 +0000 UTC m=+1414.767302202" Nov 25 10:11:58 crc kubenswrapper[4926]: I1125 10:11:58.107425 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:11:59 crc kubenswrapper[4926]: I1125 10:11:59.910217 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rcxc"] Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.092492 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6rcxc" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="registry-server" containerID="cri-o://f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4" gracePeriod=2 Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.587977 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.674570 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-utilities\") pod \"e5f64b63-b874-42b3-905a-b45833ff181b\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.674977 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54x86\" (UniqueName: \"kubernetes.io/projected/e5f64b63-b874-42b3-905a-b45833ff181b-kube-api-access-54x86\") pod \"e5f64b63-b874-42b3-905a-b45833ff181b\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.675176 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-catalog-content\") pod \"e5f64b63-b874-42b3-905a-b45833ff181b\" (UID: \"e5f64b63-b874-42b3-905a-b45833ff181b\") " Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.675438 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-utilities" (OuterVolumeSpecName: "utilities") pod "e5f64b63-b874-42b3-905a-b45833ff181b" (UID: "e5f64b63-b874-42b3-905a-b45833ff181b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.675958 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.686370 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5f64b63-b874-42b3-905a-b45833ff181b-kube-api-access-54x86" (OuterVolumeSpecName: "kube-api-access-54x86") pod "e5f64b63-b874-42b3-905a-b45833ff181b" (UID: "e5f64b63-b874-42b3-905a-b45833ff181b"). InnerVolumeSpecName "kube-api-access-54x86". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.692116 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5f64b63-b874-42b3-905a-b45833ff181b" (UID: "e5f64b63-b874-42b3-905a-b45833ff181b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.777629 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5f64b63-b874-42b3-905a-b45833ff181b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:12:01 crc kubenswrapper[4926]: I1125 10:12:01.777670 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54x86\" (UniqueName: \"kubernetes.io/projected/e5f64b63-b874-42b3-905a-b45833ff181b-kube-api-access-54x86\") on node \"crc\" DevicePath \"\"" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.102072 4926 generic.go:334] "Generic (PLEG): container finished" podID="e5f64b63-b874-42b3-905a-b45833ff181b" containerID="f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4" exitCode=0 Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.102110 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rcxc" event={"ID":"e5f64b63-b874-42b3-905a-b45833ff181b","Type":"ContainerDied","Data":"f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4"} Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.102135 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6rcxc" event={"ID":"e5f64b63-b874-42b3-905a-b45833ff181b","Type":"ContainerDied","Data":"1503606d6d49dcf385fb839a4d1caac5f7e6bddfb42b29eae604731844a2b09a"} Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.102151 4926 scope.go:117] "RemoveContainer" containerID="f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.102269 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6rcxc" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.127050 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rcxc"] Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.133562 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6rcxc"] Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.134096 4926 scope.go:117] "RemoveContainer" containerID="a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.153449 4926 scope.go:117] "RemoveContainer" containerID="928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.187330 4926 scope.go:117] "RemoveContainer" containerID="f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4" Nov 25 10:12:02 crc kubenswrapper[4926]: E1125 10:12:02.187826 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4\": container with ID starting with f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4 not found: ID does not exist" containerID="f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.187854 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4"} err="failed to get container status \"f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4\": rpc error: code = NotFound desc = could not find container \"f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4\": container with ID starting with f2e15612fd802ba4b43938c0c763a56fc8b4c9846f16bf0e60606017f70f96a4 not found: ID does not exist" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.188047 4926 scope.go:117] "RemoveContainer" containerID="a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b" Nov 25 10:12:02 crc kubenswrapper[4926]: E1125 10:12:02.188495 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b\": container with ID starting with a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b not found: ID does not exist" containerID="a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.188518 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b"} err="failed to get container status \"a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b\": rpc error: code = NotFound desc = could not find container \"a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b\": container with ID starting with a8594b1f9583b6545dcc2a9c63776cb7dc608d3859b899dbfa85efb35d3b3e1b not found: ID does not exist" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.188534 4926 scope.go:117] "RemoveContainer" containerID="928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce" Nov 25 10:12:02 crc kubenswrapper[4926]: E1125 10:12:02.188729 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce\": container with ID starting with 928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce not found: ID does not exist" containerID="928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce" Nov 25 10:12:02 crc kubenswrapper[4926]: I1125 10:12:02.188748 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce"} err="failed to get container status \"928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce\": rpc error: code = NotFound desc = could not find container \"928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce\": container with ID starting with 928912b70b3c9da552d59db97ffd1c83542b80e8df0bd9b0f3a5cac6006f0dce not found: ID does not exist" Nov 25 10:12:03 crc kubenswrapper[4926]: I1125 10:12:03.887695 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:12:03 crc kubenswrapper[4926]: I1125 10:12:03.888180 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:12:03 crc kubenswrapper[4926]: I1125 10:12:03.961666 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:12:04 crc kubenswrapper[4926]: I1125 10:12:04.024992 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" path="/var/lib/kubelet/pods/e5f64b63-b874-42b3-905a-b45833ff181b/volumes" Nov 25 10:12:04 crc kubenswrapper[4926]: I1125 10:12:04.179713 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:12:05 crc kubenswrapper[4926]: I1125 10:12:05.914380 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fwc4w"] Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.141808 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fwc4w" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="registry-server" containerID="cri-o://d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8" gracePeriod=2 Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.612914 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.671162 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-utilities\") pod \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.671402 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-catalog-content\") pod \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.671436 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k4nvb\" (UniqueName: \"kubernetes.io/projected/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-kube-api-access-k4nvb\") pod \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\" (UID: \"ae739f4b-cb8c-4f0e-9914-8736d77f25b3\") " Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.672154 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-utilities" (OuterVolumeSpecName: "utilities") pod "ae739f4b-cb8c-4f0e-9914-8736d77f25b3" (UID: "ae739f4b-cb8c-4f0e-9914-8736d77f25b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.680276 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-kube-api-access-k4nvb" (OuterVolumeSpecName: "kube-api-access-k4nvb") pod "ae739f4b-cb8c-4f0e-9914-8736d77f25b3" (UID: "ae739f4b-cb8c-4f0e-9914-8736d77f25b3"). InnerVolumeSpecName "kube-api-access-k4nvb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.723846 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ae739f4b-cb8c-4f0e-9914-8736d77f25b3" (UID: "ae739f4b-cb8c-4f0e-9914-8736d77f25b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.773887 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.773921 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k4nvb\" (UniqueName: \"kubernetes.io/projected/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-kube-api-access-k4nvb\") on node \"crc\" DevicePath \"\"" Nov 25 10:12:06 crc kubenswrapper[4926]: I1125 10:12:06.773949 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae739f4b-cb8c-4f0e-9914-8736d77f25b3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.156077 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerID="d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8" exitCode=0 Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.156174 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fwc4w" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.156165 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerDied","Data":"d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8"} Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.156667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fwc4w" event={"ID":"ae739f4b-cb8c-4f0e-9914-8736d77f25b3","Type":"ContainerDied","Data":"7384c448db31b821864e409d593d3729cf57d0d3dee16608b99129d403a33610"} Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.156699 4926 scope.go:117] "RemoveContainer" containerID="d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.182726 4926 scope.go:117] "RemoveContainer" containerID="3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.223130 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fwc4w"] Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.233166 4926 scope.go:117] "RemoveContainer" containerID="3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.241018 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fwc4w"] Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.264696 4926 scope.go:117] "RemoveContainer" containerID="d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8" Nov 25 10:12:07 crc kubenswrapper[4926]: E1125 10:12:07.265196 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8\": container with ID starting with d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8 not found: ID does not exist" containerID="d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.265238 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8"} err="failed to get container status \"d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8\": rpc error: code = NotFound desc = could not find container \"d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8\": container with ID starting with d334245f49b2319a6c35a8e57f8f894b0e60004d7ef539e1f7fa974daa3516d8 not found: ID does not exist" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.265264 4926 scope.go:117] "RemoveContainer" containerID="3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40" Nov 25 10:12:07 crc kubenswrapper[4926]: E1125 10:12:07.265868 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40\": container with ID starting with 3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40 not found: ID does not exist" containerID="3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.266044 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40"} err="failed to get container status \"3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40\": rpc error: code = NotFound desc = could not find container \"3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40\": container with ID starting with 3a9e68e4c026ee91e4601d911bcfb17f2660cbf54fee1091cbd670166fb47a40 not found: ID does not exist" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.266136 4926 scope.go:117] "RemoveContainer" containerID="3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334" Nov 25 10:12:07 crc kubenswrapper[4926]: E1125 10:12:07.266603 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334\": container with ID starting with 3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334 not found: ID does not exist" containerID="3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334" Nov 25 10:12:07 crc kubenswrapper[4926]: I1125 10:12:07.266638 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334"} err="failed to get container status \"3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334\": rpc error: code = NotFound desc = could not find container \"3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334\": container with ID starting with 3562edd23635781bcce0ee0d7b0b74e94e1e7751088cb0396ae98416be446334 not found: ID does not exist" Nov 25 10:12:08 crc kubenswrapper[4926]: I1125 10:12:08.025237 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" path="/var/lib/kubelet/pods/ae739f4b-cb8c-4f0e-9914-8736d77f25b3/volumes" Nov 25 10:12:26 crc kubenswrapper[4926]: I1125 10:12:26.558011 4926 scope.go:117] "RemoveContainer" containerID="7ba6c86f234e2a08f3839ab902b2e09e8268a1369b066b66413dd1fbc6f64bbb" Nov 25 10:12:26 crc kubenswrapper[4926]: I1125 10:12:26.606096 4926 scope.go:117] "RemoveContainer" containerID="07893fd6ab086c07011aedb868a2d874eb2ee0c6229d38efb7780171a6b1439b" Nov 25 10:12:44 crc kubenswrapper[4926]: I1125 10:12:44.021605 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:12:44 crc kubenswrapper[4926]: I1125 10:12:44.022171 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:13:14 crc kubenswrapper[4926]: I1125 10:13:14.017554 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:13:14 crc kubenswrapper[4926]: I1125 10:13:14.018273 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.024561 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.025422 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.025588 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.026386 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.026477 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" gracePeriod=600 Nov 25 10:13:44 crc kubenswrapper[4926]: E1125 10:13:44.152221 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.170670 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" exitCode=0 Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.170719 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625"} Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.170756 4926 scope.go:117] "RemoveContainer" containerID="780e5b5fae7c7161700dd19e5dd8229ab43351e7385e07fc9c5c1c4ee3eb5f66" Nov 25 10:13:44 crc kubenswrapper[4926]: I1125 10:13:44.171367 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:13:44 crc kubenswrapper[4926]: E1125 10:13:44.171703 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:13:51 crc kubenswrapper[4926]: I1125 10:13:51.237473 4926 generic.go:334] "Generic (PLEG): container finished" podID="ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" containerID="4ddbb6fda74711116d29048e3312ec8c3d69fd0ae587fe2f7168188a0c46dead" exitCode=0 Nov 25 10:13:51 crc kubenswrapper[4926]: I1125 10:13:51.237562 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" event={"ID":"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9","Type":"ContainerDied","Data":"4ddbb6fda74711116d29048e3312ec8c3d69fd0ae587fe2f7168188a0c46dead"} Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.609118 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.715950 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-bootstrap-combined-ca-bundle\") pod \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.716277 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cwc4r\" (UniqueName: \"kubernetes.io/projected/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-kube-api-access-cwc4r\") pod \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.716322 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-inventory\") pod \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.716369 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-ssh-key\") pod \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\" (UID: \"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9\") " Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.721375 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-kube-api-access-cwc4r" (OuterVolumeSpecName: "kube-api-access-cwc4r") pod "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" (UID: "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9"). InnerVolumeSpecName "kube-api-access-cwc4r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.721596 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" (UID: "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.740391 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-inventory" (OuterVolumeSpecName: "inventory") pod "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" (UID: "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.741214 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" (UID: "ac1288a0-8cc2-49b5-b9d4-f0402c00aba9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.818833 4926 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.818873 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cwc4r\" (UniqueName: \"kubernetes.io/projected/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-kube-api-access-cwc4r\") on node \"crc\" DevicePath \"\"" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.818889 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:13:52 crc kubenswrapper[4926]: I1125 10:13:52.818902 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac1288a0-8cc2-49b5-b9d4-f0402c00aba9-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.259178 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" event={"ID":"ac1288a0-8cc2-49b5-b9d4-f0402c00aba9","Type":"ContainerDied","Data":"8b8b184198a4af75d39dcdd805871d7644a80d52a2061d4d773b478b2c48a595"} Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.259227 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b8b184198a4af75d39dcdd805871d7644a80d52a2061d4d773b478b2c48a595" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.259247 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367271 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf"] Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367709 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="registry-server" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367733 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="registry-server" Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367763 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="extract-utilities" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367770 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="extract-utilities" Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367784 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="extract-content" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367792 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="extract-content" Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367803 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="extract-utilities" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367812 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="extract-utilities" Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367831 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367841 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367852 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="extract-content" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367866 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="extract-content" Nov 25 10:13:53 crc kubenswrapper[4926]: E1125 10:13:53.367882 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="registry-server" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.367893 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="registry-server" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.368130 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5f64b63-b874-42b3-905a-b45833ff181b" containerName="registry-server" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.368147 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1288a0-8cc2-49b5-b9d4-f0402c00aba9" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.368162 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae739f4b-cb8c-4f0e-9914-8736d77f25b3" containerName="registry-server" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.368837 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.375277 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.375586 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.375663 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.375684 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.388450 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf"] Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.427830 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/0e9ec79e-2dba-4832-baa4-8b630738f623-kube-api-access-xqkk5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.427914 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.428057 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.529998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/0e9ec79e-2dba-4832-baa4-8b630738f623-kube-api-access-xqkk5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.530090 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.530118 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.534708 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.534931 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.561964 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/0e9ec79e-2dba-4832-baa4-8b630738f623-kube-api-access-xqkk5\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-jsngf\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:53 crc kubenswrapper[4926]: I1125 10:13:53.703753 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:13:54 crc kubenswrapper[4926]: I1125 10:13:54.235754 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf"] Nov 25 10:13:54 crc kubenswrapper[4926]: I1125 10:13:54.268575 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" event={"ID":"0e9ec79e-2dba-4832-baa4-8b630738f623","Type":"ContainerStarted","Data":"cc49a5968abe2f9899122a55a24d8d0a2e47a0cd189a8f3b28a3ad1d482fb656"} Nov 25 10:13:55 crc kubenswrapper[4926]: I1125 10:13:55.013074 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:13:55 crc kubenswrapper[4926]: E1125 10:13:55.013586 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:13:55 crc kubenswrapper[4926]: I1125 10:13:55.283839 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" event={"ID":"0e9ec79e-2dba-4832-baa4-8b630738f623","Type":"ContainerStarted","Data":"4aa61863f0724780900bcdecf43c018ba7dd752c12e3c0926f793dc10991affb"} Nov 25 10:13:55 crc kubenswrapper[4926]: I1125 10:13:55.309270 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" podStartSLOduration=1.868830831 podStartE2EDuration="2.309236045s" podCreationTimestamp="2025-11-25 10:13:53 +0000 UTC" firstStartedPulling="2025-11-25 10:13:54.243294267 +0000 UTC m=+1530.927250684" lastFinishedPulling="2025-11-25 10:13:54.683699431 +0000 UTC m=+1531.367655898" observedRunningTime="2025-11-25 10:13:55.299336059 +0000 UTC m=+1531.983292496" watchObservedRunningTime="2025-11-25 10:13:55.309236045 +0000 UTC m=+1531.993192502" Nov 25 10:14:08 crc kubenswrapper[4926]: I1125 10:14:08.012694 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:14:08 crc kubenswrapper[4926]: E1125 10:14:08.013752 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:14:20 crc kubenswrapper[4926]: I1125 10:14:20.013226 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:14:20 crc kubenswrapper[4926]: E1125 10:14:20.014159 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:14:26 crc kubenswrapper[4926]: I1125 10:14:26.753209 4926 scope.go:117] "RemoveContainer" containerID="acdc94c2d20d7b8c6619f73c327ef238d106843c43f5bb29fe9ba111c385f3f3" Nov 25 10:14:26 crc kubenswrapper[4926]: I1125 10:14:26.776484 4926 scope.go:117] "RemoveContainer" containerID="5394837af21313f52061f9a04937d79fc8f150ca635e195524a8318d2224d722" Nov 25 10:14:33 crc kubenswrapper[4926]: I1125 10:14:33.012056 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:14:33 crc kubenswrapper[4926]: E1125 10:14:33.012733 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:14:46 crc kubenswrapper[4926]: I1125 10:14:46.012978 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:14:46 crc kubenswrapper[4926]: E1125 10:14:46.013645 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.145291 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx"] Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.147007 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.149562 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.150067 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.163006 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx"] Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.292658 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45d4f8be-0e15-4b98-9348-75cc07786909-config-volume\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.292710 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45d4f8be-0e15-4b98-9348-75cc07786909-secret-volume\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.292913 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmd6r\" (UniqueName: \"kubernetes.io/projected/45d4f8be-0e15-4b98-9348-75cc07786909-kube-api-access-bmd6r\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.394679 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmd6r\" (UniqueName: \"kubernetes.io/projected/45d4f8be-0e15-4b98-9348-75cc07786909-kube-api-access-bmd6r\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.394979 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45d4f8be-0e15-4b98-9348-75cc07786909-config-volume\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.395080 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45d4f8be-0e15-4b98-9348-75cc07786909-secret-volume\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.395885 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45d4f8be-0e15-4b98-9348-75cc07786909-config-volume\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.405971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45d4f8be-0e15-4b98-9348-75cc07786909-secret-volume\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.410511 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmd6r\" (UniqueName: \"kubernetes.io/projected/45d4f8be-0e15-4b98-9348-75cc07786909-kube-api-access-bmd6r\") pod \"collect-profiles-29401095-s8ztx\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.470421 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:00 crc kubenswrapper[4926]: I1125 10:15:00.934486 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx"] Nov 25 10:15:01 crc kubenswrapper[4926]: I1125 10:15:01.012599 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:15:01 crc kubenswrapper[4926]: E1125 10:15:01.013478 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:15:01 crc kubenswrapper[4926]: I1125 10:15:01.944826 4926 generic.go:334] "Generic (PLEG): container finished" podID="45d4f8be-0e15-4b98-9348-75cc07786909" containerID="078b654e596fa8f735cab40c07d9b2f5688bffb021f20d0eee8f8f62fc8147f9" exitCode=0 Nov 25 10:15:01 crc kubenswrapper[4926]: I1125 10:15:01.944892 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" event={"ID":"45d4f8be-0e15-4b98-9348-75cc07786909","Type":"ContainerDied","Data":"078b654e596fa8f735cab40c07d9b2f5688bffb021f20d0eee8f8f62fc8147f9"} Nov 25 10:15:01 crc kubenswrapper[4926]: I1125 10:15:01.945203 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" event={"ID":"45d4f8be-0e15-4b98-9348-75cc07786909","Type":"ContainerStarted","Data":"0e6b49e236ad9ae3df8624913ddcf8fa61e77b7b559c114bfd2614fdba2fef49"} Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.282197 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.456856 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45d4f8be-0e15-4b98-9348-75cc07786909-config-volume\") pod \"45d4f8be-0e15-4b98-9348-75cc07786909\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.457153 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45d4f8be-0e15-4b98-9348-75cc07786909-secret-volume\") pod \"45d4f8be-0e15-4b98-9348-75cc07786909\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.457210 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmd6r\" (UniqueName: \"kubernetes.io/projected/45d4f8be-0e15-4b98-9348-75cc07786909-kube-api-access-bmd6r\") pod \"45d4f8be-0e15-4b98-9348-75cc07786909\" (UID: \"45d4f8be-0e15-4b98-9348-75cc07786909\") " Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.460474 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45d4f8be-0e15-4b98-9348-75cc07786909-config-volume" (OuterVolumeSpecName: "config-volume") pod "45d4f8be-0e15-4b98-9348-75cc07786909" (UID: "45d4f8be-0e15-4b98-9348-75cc07786909"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.462457 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45d4f8be-0e15-4b98-9348-75cc07786909-kube-api-access-bmd6r" (OuterVolumeSpecName: "kube-api-access-bmd6r") pod "45d4f8be-0e15-4b98-9348-75cc07786909" (UID: "45d4f8be-0e15-4b98-9348-75cc07786909"). InnerVolumeSpecName "kube-api-access-bmd6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.473214 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45d4f8be-0e15-4b98-9348-75cc07786909-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "45d4f8be-0e15-4b98-9348-75cc07786909" (UID: "45d4f8be-0e15-4b98-9348-75cc07786909"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.558722 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/45d4f8be-0e15-4b98-9348-75cc07786909-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.558752 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmd6r\" (UniqueName: \"kubernetes.io/projected/45d4f8be-0e15-4b98-9348-75cc07786909-kube-api-access-bmd6r\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.558760 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/45d4f8be-0e15-4b98-9348-75cc07786909-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.966784 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" event={"ID":"45d4f8be-0e15-4b98-9348-75cc07786909","Type":"ContainerDied","Data":"0e6b49e236ad9ae3df8624913ddcf8fa61e77b7b559c114bfd2614fdba2fef49"} Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.966825 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e6b49e236ad9ae3df8624913ddcf8fa61e77b7b559c114bfd2614fdba2fef49" Nov 25 10:15:03 crc kubenswrapper[4926]: I1125 10:15:03.966860 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401095-s8ztx" Nov 25 10:15:09 crc kubenswrapper[4926]: I1125 10:15:09.006192 4926 generic.go:334] "Generic (PLEG): container finished" podID="0e9ec79e-2dba-4832-baa4-8b630738f623" containerID="4aa61863f0724780900bcdecf43c018ba7dd752c12e3c0926f793dc10991affb" exitCode=0 Nov 25 10:15:09 crc kubenswrapper[4926]: I1125 10:15:09.006268 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" event={"ID":"0e9ec79e-2dba-4832-baa4-8b630738f623","Type":"ContainerDied","Data":"4aa61863f0724780900bcdecf43c018ba7dd752c12e3c0926f793dc10991affb"} Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.381280 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.481693 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-ssh-key\") pod \"0e9ec79e-2dba-4832-baa4-8b630738f623\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.481842 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-inventory\") pod \"0e9ec79e-2dba-4832-baa4-8b630738f623\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.482025 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/0e9ec79e-2dba-4832-baa4-8b630738f623-kube-api-access-xqkk5\") pod \"0e9ec79e-2dba-4832-baa4-8b630738f623\" (UID: \"0e9ec79e-2dba-4832-baa4-8b630738f623\") " Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.487975 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e9ec79e-2dba-4832-baa4-8b630738f623-kube-api-access-xqkk5" (OuterVolumeSpecName: "kube-api-access-xqkk5") pod "0e9ec79e-2dba-4832-baa4-8b630738f623" (UID: "0e9ec79e-2dba-4832-baa4-8b630738f623"). InnerVolumeSpecName "kube-api-access-xqkk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.511100 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0e9ec79e-2dba-4832-baa4-8b630738f623" (UID: "0e9ec79e-2dba-4832-baa4-8b630738f623"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.511242 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-inventory" (OuterVolumeSpecName: "inventory") pod "0e9ec79e-2dba-4832-baa4-8b630738f623" (UID: "0e9ec79e-2dba-4832-baa4-8b630738f623"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.585353 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.585397 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e9ec79e-2dba-4832-baa4-8b630738f623-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:10 crc kubenswrapper[4926]: I1125 10:15:10.585411 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqkk5\" (UniqueName: \"kubernetes.io/projected/0e9ec79e-2dba-4832-baa4-8b630738f623-kube-api-access-xqkk5\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.025767 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" event={"ID":"0e9ec79e-2dba-4832-baa4-8b630738f623","Type":"ContainerDied","Data":"cc49a5968abe2f9899122a55a24d8d0a2e47a0cd189a8f3b28a3ad1d482fb656"} Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.026114 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc49a5968abe2f9899122a55a24d8d0a2e47a0cd189a8f3b28a3ad1d482fb656" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.025840 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-jsngf" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.152083 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8"] Nov 25 10:15:11 crc kubenswrapper[4926]: E1125 10:15:11.152504 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e9ec79e-2dba-4832-baa4-8b630738f623" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.152535 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e9ec79e-2dba-4832-baa4-8b630738f623" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:11 crc kubenswrapper[4926]: E1125 10:15:11.152600 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45d4f8be-0e15-4b98-9348-75cc07786909" containerName="collect-profiles" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.152612 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="45d4f8be-0e15-4b98-9348-75cc07786909" containerName="collect-profiles" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.152836 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="45d4f8be-0e15-4b98-9348-75cc07786909" containerName="collect-profiles" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.152864 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e9ec79e-2dba-4832-baa4-8b630738f623" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.153603 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.158549 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.158795 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.159033 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.159240 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.166545 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8"] Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.225887 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmg9c\" (UniqueName: \"kubernetes.io/projected/2f83a69b-6808-4409-b29a-72c78653c714-kube-api-access-gmg9c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.225972 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.226095 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.327730 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmg9c\" (UniqueName: \"kubernetes.io/projected/2f83a69b-6808-4409-b29a-72c78653c714-kube-api-access-gmg9c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.327784 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.327817 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.334322 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.334425 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.349901 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmg9c\" (UniqueName: \"kubernetes.io/projected/2f83a69b-6808-4409-b29a-72c78653c714-kube-api-access-gmg9c\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-lkln8\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.479707 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:11 crc kubenswrapper[4926]: I1125 10:15:11.987560 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8"] Nov 25 10:15:12 crc kubenswrapper[4926]: I1125 10:15:12.040731 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" event={"ID":"2f83a69b-6808-4409-b29a-72c78653c714","Type":"ContainerStarted","Data":"ffec7e6c48b19aeafcd0951b72a48dc108401479f472f5fae90174227bcff7fe"} Nov 25 10:15:13 crc kubenswrapper[4926]: I1125 10:15:13.050881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" event={"ID":"2f83a69b-6808-4409-b29a-72c78653c714","Type":"ContainerStarted","Data":"c943fb82e3bb748b734a347d0b1cae7599ab5abe0b2ca949a1249fa5c52dd0a5"} Nov 25 10:15:15 crc kubenswrapper[4926]: I1125 10:15:15.012379 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:15:15 crc kubenswrapper[4926]: E1125 10:15:15.012823 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:15:17 crc kubenswrapper[4926]: I1125 10:15:17.053056 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" podStartSLOduration=5.628987778 podStartE2EDuration="6.053034965s" podCreationTimestamp="2025-11-25 10:15:11 +0000 UTC" firstStartedPulling="2025-11-25 10:15:11.998483672 +0000 UTC m=+1608.682440089" lastFinishedPulling="2025-11-25 10:15:12.422530849 +0000 UTC m=+1609.106487276" observedRunningTime="2025-11-25 10:15:13.073913896 +0000 UTC m=+1609.757870323" watchObservedRunningTime="2025-11-25 10:15:17.053034965 +0000 UTC m=+1613.736991402" Nov 25 10:15:17 crc kubenswrapper[4926]: I1125 10:15:17.055478 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-298a-account-create-fpwbz"] Nov 25 10:15:17 crc kubenswrapper[4926]: I1125 10:15:17.066219 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-pmc8k"] Nov 25 10:15:17 crc kubenswrapper[4926]: I1125 10:15:17.082075 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-298a-account-create-fpwbz"] Nov 25 10:15:17 crc kubenswrapper[4926]: I1125 10:15:17.092914 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-pmc8k"] Nov 25 10:15:18 crc kubenswrapper[4926]: I1125 10:15:18.025465 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="230dd925-418b-4bc1-b62f-b70befdd2a59" path="/var/lib/kubelet/pods/230dd925-418b-4bc1-b62f-b70befdd2a59/volumes" Nov 25 10:15:18 crc kubenswrapper[4926]: I1125 10:15:18.026692 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="479a95d6-7f72-4322-929a-d124f7722c40" path="/var/lib/kubelet/pods/479a95d6-7f72-4322-929a-d124f7722c40/volumes" Nov 25 10:15:18 crc kubenswrapper[4926]: I1125 10:15:18.098293 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f83a69b-6808-4409-b29a-72c78653c714" containerID="c943fb82e3bb748b734a347d0b1cae7599ab5abe0b2ca949a1249fa5c52dd0a5" exitCode=0 Nov 25 10:15:18 crc kubenswrapper[4926]: I1125 10:15:18.098334 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" event={"ID":"2f83a69b-6808-4409-b29a-72c78653c714","Type":"ContainerDied","Data":"c943fb82e3bb748b734a347d0b1cae7599ab5abe0b2ca949a1249fa5c52dd0a5"} Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.472655 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.587280 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmg9c\" (UniqueName: \"kubernetes.io/projected/2f83a69b-6808-4409-b29a-72c78653c714-kube-api-access-gmg9c\") pod \"2f83a69b-6808-4409-b29a-72c78653c714\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.587382 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-ssh-key\") pod \"2f83a69b-6808-4409-b29a-72c78653c714\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.587456 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-inventory\") pod \"2f83a69b-6808-4409-b29a-72c78653c714\" (UID: \"2f83a69b-6808-4409-b29a-72c78653c714\") " Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.598215 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f83a69b-6808-4409-b29a-72c78653c714-kube-api-access-gmg9c" (OuterVolumeSpecName: "kube-api-access-gmg9c") pod "2f83a69b-6808-4409-b29a-72c78653c714" (UID: "2f83a69b-6808-4409-b29a-72c78653c714"). InnerVolumeSpecName "kube-api-access-gmg9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.612639 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2f83a69b-6808-4409-b29a-72c78653c714" (UID: "2f83a69b-6808-4409-b29a-72c78653c714"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.617755 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-inventory" (OuterVolumeSpecName: "inventory") pod "2f83a69b-6808-4409-b29a-72c78653c714" (UID: "2f83a69b-6808-4409-b29a-72c78653c714"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.692546 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmg9c\" (UniqueName: \"kubernetes.io/projected/2f83a69b-6808-4409-b29a-72c78653c714-kube-api-access-gmg9c\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.692631 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:19 crc kubenswrapper[4926]: I1125 10:15:19.692651 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2f83a69b-6808-4409-b29a-72c78653c714-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.122641 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" event={"ID":"2f83a69b-6808-4409-b29a-72c78653c714","Type":"ContainerDied","Data":"ffec7e6c48b19aeafcd0951b72a48dc108401479f472f5fae90174227bcff7fe"} Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.122734 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-lkln8" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.122748 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffec7e6c48b19aeafcd0951b72a48dc108401479f472f5fae90174227bcff7fe" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.199789 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5"] Nov 25 10:15:20 crc kubenswrapper[4926]: E1125 10:15:20.200208 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f83a69b-6808-4409-b29a-72c78653c714" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.200234 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f83a69b-6808-4409-b29a-72c78653c714" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.200439 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f83a69b-6808-4409-b29a-72c78653c714" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.201124 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.203275 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.203655 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.203841 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.204015 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.211735 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5"] Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.304202 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.304248 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr6tj\" (UniqueName: \"kubernetes.io/projected/311b8250-d497-4cf0-9b38-72e00d16f422-kube-api-access-lr6tj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.304457 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.405779 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.405904 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.405933 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr6tj\" (UniqueName: \"kubernetes.io/projected/311b8250-d497-4cf0-9b38-72e00d16f422-kube-api-access-lr6tj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.410570 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.410615 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.423380 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr6tj\" (UniqueName: \"kubernetes.io/projected/311b8250-d497-4cf0-9b38-72e00d16f422-kube-api-access-lr6tj\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-skbf5\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:20 crc kubenswrapper[4926]: I1125 10:15:20.563836 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:21 crc kubenswrapper[4926]: I1125 10:15:21.029380 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9vf4k"] Nov 25 10:15:21 crc kubenswrapper[4926]: I1125 10:15:21.036630 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9vf4k"] Nov 25 10:15:21 crc kubenswrapper[4926]: I1125 10:15:21.121468 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5"] Nov 25 10:15:21 crc kubenswrapper[4926]: I1125 10:15:21.124896 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:15:21 crc kubenswrapper[4926]: I1125 10:15:21.135334 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" event={"ID":"311b8250-d497-4cf0-9b38-72e00d16f422","Type":"ContainerStarted","Data":"ddcf83ddc60f4fb434e3b0dca3c79a0ca0017e5e75fcdfd26a63d972120ab97a"} Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.023220 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0" path="/var/lib/kubelet/pods/c48a5eda-1ab4-41d5-9a82-0b6e22a2bef0/volumes" Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.027064 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-fa79-account-create-c9vvp"] Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.037819 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-b544s"] Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.046483 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-b21d-account-create-cbzxv"] Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.053615 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-b21d-account-create-cbzxv"] Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.059596 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-fa79-account-create-c9vvp"] Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.065419 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-b544s"] Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.144881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" event={"ID":"311b8250-d497-4cf0-9b38-72e00d16f422","Type":"ContainerStarted","Data":"0e6d716ccdf8b687f57c70c7baafe061a553f4379454b2c6e4f6bb7d450013ad"} Nov 25 10:15:22 crc kubenswrapper[4926]: I1125 10:15:22.166575 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" podStartSLOduration=1.665258264 podStartE2EDuration="2.166557522s" podCreationTimestamp="2025-11-25 10:15:20 +0000 UTC" firstStartedPulling="2025-11-25 10:15:21.12458445 +0000 UTC m=+1617.808540877" lastFinishedPulling="2025-11-25 10:15:21.625883688 +0000 UTC m=+1618.309840135" observedRunningTime="2025-11-25 10:15:22.162371678 +0000 UTC m=+1618.846328105" watchObservedRunningTime="2025-11-25 10:15:22.166557522 +0000 UTC m=+1618.850513949" Nov 25 10:15:24 crc kubenswrapper[4926]: I1125 10:15:24.025288 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="109f1cf1-f016-438b-b2b9-2eaabf3ba1c9" path="/var/lib/kubelet/pods/109f1cf1-f016-438b-b2b9-2eaabf3ba1c9/volumes" Nov 25 10:15:24 crc kubenswrapper[4926]: I1125 10:15:24.026170 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b24c1a1f-5324-4699-90a5-b539c855acb6" path="/var/lib/kubelet/pods/b24c1a1f-5324-4699-90a5-b539c855acb6/volumes" Nov 25 10:15:24 crc kubenswrapper[4926]: I1125 10:15:24.026740 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2bac7de-ff6d-4b5f-85c0-c3653855f844" path="/var/lib/kubelet/pods/f2bac7de-ff6d-4b5f-85c0-c3653855f844/volumes" Nov 25 10:15:26 crc kubenswrapper[4926]: I1125 10:15:26.839757 4926 scope.go:117] "RemoveContainer" containerID="e8e58cf43c7ecf87b491b558284ddfb60380a8ba9ca0087dac53819159571b5b" Nov 25 10:15:26 crc kubenswrapper[4926]: I1125 10:15:26.885541 4926 scope.go:117] "RemoveContainer" containerID="72f9706efbd33dcb0f06a30f151666de3d9ea9209b67403a4ea7d476080f813c" Nov 25 10:15:26 crc kubenswrapper[4926]: I1125 10:15:26.917484 4926 scope.go:117] "RemoveContainer" containerID="6b1ced5690cceae03aea9163438b50de336a67bdbc21f305a8a4946b9a26fcc0" Nov 25 10:15:26 crc kubenswrapper[4926]: I1125 10:15:26.958394 4926 scope.go:117] "RemoveContainer" containerID="6a6c361b7ee2665024d77674ec3b601fa009b3f406f54f5843f44f46f20a1711" Nov 25 10:15:26 crc kubenswrapper[4926]: I1125 10:15:26.999683 4926 scope.go:117] "RemoveContainer" containerID="df8e3130e3f3beaa8b38fdb49b0e47a61cb1d15fc72b2c5aacc1df874dfbe2b4" Nov 25 10:15:27 crc kubenswrapper[4926]: I1125 10:15:27.042676 4926 scope.go:117] "RemoveContainer" containerID="e08ffb87f1516f45eb4fb713ffb56a32d79a6a754fb21099998e7f5b59637f67" Nov 25 10:15:29 crc kubenswrapper[4926]: I1125 10:15:29.012437 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:15:29 crc kubenswrapper[4926]: E1125 10:15:29.012928 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:15:40 crc kubenswrapper[4926]: I1125 10:15:40.013626 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:15:40 crc kubenswrapper[4926]: E1125 10:15:40.014449 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:15:40 crc kubenswrapper[4926]: I1125 10:15:40.039010 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-grrv7"] Nov 25 10:15:40 crc kubenswrapper[4926]: I1125 10:15:40.050458 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-grrv7"] Nov 25 10:15:42 crc kubenswrapper[4926]: I1125 10:15:42.025236 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ceab1d9b-89a4-412c-b720-b46193249d81" path="/var/lib/kubelet/pods/ceab1d9b-89a4-412c-b720-b46193249d81/volumes" Nov 25 10:15:51 crc kubenswrapper[4926]: I1125 10:15:51.013778 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:15:51 crc kubenswrapper[4926]: E1125 10:15:51.014651 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.042368 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-88zw6"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.052097 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-9d9f-account-create-n7kx2"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.062919 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-ea5b-account-create-4lqsw"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.069684 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-d6aa-account-create-ksf77"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.077308 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-jvjxm"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.088299 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-nrx5l"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.095268 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-ea5b-account-create-4lqsw"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.103827 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-88zw6"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.112337 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-9d9f-account-create-n7kx2"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.120575 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-d6aa-account-create-ksf77"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.129088 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-jvjxm"] Nov 25 10:15:52 crc kubenswrapper[4926]: I1125 10:15:52.136152 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-nrx5l"] Nov 25 10:15:54 crc kubenswrapper[4926]: I1125 10:15:54.028397 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09f50a9c-beb0-4c48-9081-b740a6a0fc60" path="/var/lib/kubelet/pods/09f50a9c-beb0-4c48-9081-b740a6a0fc60/volumes" Nov 25 10:15:54 crc kubenswrapper[4926]: I1125 10:15:54.029281 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a65fcf0-3c1c-4832-a039-ddd082b5c7bd" path="/var/lib/kubelet/pods/8a65fcf0-3c1c-4832-a039-ddd082b5c7bd/volumes" Nov 25 10:15:54 crc kubenswrapper[4926]: I1125 10:15:54.029994 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9a07ca9-335b-484d-ba61-d9b4f798dd98" path="/var/lib/kubelet/pods/d9a07ca9-335b-484d-ba61-d9b4f798dd98/volumes" Nov 25 10:15:54 crc kubenswrapper[4926]: I1125 10:15:54.030670 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1571b13-1388-4ce6-b85d-c3250984ea99" path="/var/lib/kubelet/pods/e1571b13-1388-4ce6-b85d-c3250984ea99/volumes" Nov 25 10:15:54 crc kubenswrapper[4926]: I1125 10:15:54.031911 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecbca8cc-2f37-49d7-8038-a20c2d68e0eb" path="/var/lib/kubelet/pods/ecbca8cc-2f37-49d7-8038-a20c2d68e0eb/volumes" Nov 25 10:15:54 crc kubenswrapper[4926]: I1125 10:15:54.032548 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f66b61a5-3157-42ea-9dad-6b112c268960" path="/var/lib/kubelet/pods/f66b61a5-3157-42ea-9dad-6b112c268960/volumes" Nov 25 10:15:56 crc kubenswrapper[4926]: I1125 10:15:56.031041 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-c66wr"] Nov 25 10:15:56 crc kubenswrapper[4926]: I1125 10:15:56.038129 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-c66wr"] Nov 25 10:15:57 crc kubenswrapper[4926]: I1125 10:15:57.448990 4926 generic.go:334] "Generic (PLEG): container finished" podID="311b8250-d497-4cf0-9b38-72e00d16f422" containerID="0e6d716ccdf8b687f57c70c7baafe061a553f4379454b2c6e4f6bb7d450013ad" exitCode=0 Nov 25 10:15:57 crc kubenswrapper[4926]: I1125 10:15:57.449234 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" event={"ID":"311b8250-d497-4cf0-9b38-72e00d16f422","Type":"ContainerDied","Data":"0e6d716ccdf8b687f57c70c7baafe061a553f4379454b2c6e4f6bb7d450013ad"} Nov 25 10:15:58 crc kubenswrapper[4926]: I1125 10:15:58.023644 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25d7aade-38a4-438b-a75d-420103b7bcdc" path="/var/lib/kubelet/pods/25d7aade-38a4-438b-a75d-420103b7bcdc/volumes" Nov 25 10:15:58 crc kubenswrapper[4926]: I1125 10:15:58.871265 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.068185 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-ssh-key\") pod \"311b8250-d497-4cf0-9b38-72e00d16f422\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.069679 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr6tj\" (UniqueName: \"kubernetes.io/projected/311b8250-d497-4cf0-9b38-72e00d16f422-kube-api-access-lr6tj\") pod \"311b8250-d497-4cf0-9b38-72e00d16f422\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.070052 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-inventory\") pod \"311b8250-d497-4cf0-9b38-72e00d16f422\" (UID: \"311b8250-d497-4cf0-9b38-72e00d16f422\") " Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.076099 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/311b8250-d497-4cf0-9b38-72e00d16f422-kube-api-access-lr6tj" (OuterVolumeSpecName: "kube-api-access-lr6tj") pod "311b8250-d497-4cf0-9b38-72e00d16f422" (UID: "311b8250-d497-4cf0-9b38-72e00d16f422"). InnerVolumeSpecName "kube-api-access-lr6tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.108442 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-inventory" (OuterVolumeSpecName: "inventory") pod "311b8250-d497-4cf0-9b38-72e00d16f422" (UID: "311b8250-d497-4cf0-9b38-72e00d16f422"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.110538 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "311b8250-d497-4cf0-9b38-72e00d16f422" (UID: "311b8250-d497-4cf0-9b38-72e00d16f422"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.173258 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.173326 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr6tj\" (UniqueName: \"kubernetes.io/projected/311b8250-d497-4cf0-9b38-72e00d16f422-kube-api-access-lr6tj\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.173343 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/311b8250-d497-4cf0-9b38-72e00d16f422-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.469822 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" event={"ID":"311b8250-d497-4cf0-9b38-72e00d16f422","Type":"ContainerDied","Data":"ddcf83ddc60f4fb434e3b0dca3c79a0ca0017e5e75fcdfd26a63d972120ab97a"} Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.469865 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ddcf83ddc60f4fb434e3b0dca3c79a0ca0017e5e75fcdfd26a63d972120ab97a" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.469867 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-skbf5" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.534410 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v"] Nov 25 10:15:59 crc kubenswrapper[4926]: E1125 10:15:59.534766 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="311b8250-d497-4cf0-9b38-72e00d16f422" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.534784 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="311b8250-d497-4cf0-9b38-72e00d16f422" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.534996 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="311b8250-d497-4cf0-9b38-72e00d16f422" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.535539 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.539741 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.539759 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.539875 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.539957 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.553429 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v"] Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.681765 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.681856 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-692q6\" (UniqueName: \"kubernetes.io/projected/74f34995-35ea-4875-9a69-61196081a80f-kube-api-access-692q6\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.681985 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.783700 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.783802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.783887 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-692q6\" (UniqueName: \"kubernetes.io/projected/74f34995-35ea-4875-9a69-61196081a80f-kube-api-access-692q6\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.788229 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.788553 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.799777 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-692q6\" (UniqueName: \"kubernetes.io/projected/74f34995-35ea-4875-9a69-61196081a80f-kube-api-access-692q6\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:15:59 crc kubenswrapper[4926]: I1125 10:15:59.858435 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:16:00 crc kubenswrapper[4926]: I1125 10:16:00.339822 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v"] Nov 25 10:16:00 crc kubenswrapper[4926]: I1125 10:16:00.479711 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" event={"ID":"74f34995-35ea-4875-9a69-61196081a80f","Type":"ContainerStarted","Data":"fb2ddd784c710b0f08698d1f7a211e01eba046d9bef11bf11ae163f2afa9f6da"} Nov 25 10:16:01 crc kubenswrapper[4926]: I1125 10:16:01.487437 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" event={"ID":"74f34995-35ea-4875-9a69-61196081a80f","Type":"ContainerStarted","Data":"4ef5962a54f6b80da5cdb85e5e7b9d85dbfc2d7b2b9895090abb55b95567b2b1"} Nov 25 10:16:01 crc kubenswrapper[4926]: I1125 10:16:01.509848 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" podStartSLOduration=1.804656907 podStartE2EDuration="2.509831097s" podCreationTimestamp="2025-11-25 10:15:59 +0000 UTC" firstStartedPulling="2025-11-25 10:16:00.345594759 +0000 UTC m=+1657.029551176" lastFinishedPulling="2025-11-25 10:16:01.050768919 +0000 UTC m=+1657.734725366" observedRunningTime="2025-11-25 10:16:01.502373062 +0000 UTC m=+1658.186329489" watchObservedRunningTime="2025-11-25 10:16:01.509831097 +0000 UTC m=+1658.193787524" Nov 25 10:16:02 crc kubenswrapper[4926]: I1125 10:16:02.013580 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:16:02 crc kubenswrapper[4926]: E1125 10:16:02.013807 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:16:05 crc kubenswrapper[4926]: I1125 10:16:05.525140 4926 generic.go:334] "Generic (PLEG): container finished" podID="74f34995-35ea-4875-9a69-61196081a80f" containerID="4ef5962a54f6b80da5cdb85e5e7b9d85dbfc2d7b2b9895090abb55b95567b2b1" exitCode=0 Nov 25 10:16:05 crc kubenswrapper[4926]: I1125 10:16:05.525231 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" event={"ID":"74f34995-35ea-4875-9a69-61196081a80f","Type":"ContainerDied","Data":"4ef5962a54f6b80da5cdb85e5e7b9d85dbfc2d7b2b9895090abb55b95567b2b1"} Nov 25 10:16:06 crc kubenswrapper[4926]: I1125 10:16:06.931583 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.088346 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-692q6\" (UniqueName: \"kubernetes.io/projected/74f34995-35ea-4875-9a69-61196081a80f-kube-api-access-692q6\") pod \"74f34995-35ea-4875-9a69-61196081a80f\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.088494 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-ssh-key\") pod \"74f34995-35ea-4875-9a69-61196081a80f\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.088531 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-inventory\") pod \"74f34995-35ea-4875-9a69-61196081a80f\" (UID: \"74f34995-35ea-4875-9a69-61196081a80f\") " Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.100366 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74f34995-35ea-4875-9a69-61196081a80f-kube-api-access-692q6" (OuterVolumeSpecName: "kube-api-access-692q6") pod "74f34995-35ea-4875-9a69-61196081a80f" (UID: "74f34995-35ea-4875-9a69-61196081a80f"). InnerVolumeSpecName "kube-api-access-692q6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.115829 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "74f34995-35ea-4875-9a69-61196081a80f" (UID: "74f34995-35ea-4875-9a69-61196081a80f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.119100 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-inventory" (OuterVolumeSpecName: "inventory") pod "74f34995-35ea-4875-9a69-61196081a80f" (UID: "74f34995-35ea-4875-9a69-61196081a80f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.191249 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-692q6\" (UniqueName: \"kubernetes.io/projected/74f34995-35ea-4875-9a69-61196081a80f-kube-api-access-692q6\") on node \"crc\" DevicePath \"\"" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.191277 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.191286 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/74f34995-35ea-4875-9a69-61196081a80f-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.543664 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" event={"ID":"74f34995-35ea-4875-9a69-61196081a80f","Type":"ContainerDied","Data":"fb2ddd784c710b0f08698d1f7a211e01eba046d9bef11bf11ae163f2afa9f6da"} Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.543703 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.543712 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb2ddd784c710b0f08698d1f7a211e01eba046d9bef11bf11ae163f2afa9f6da" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.604787 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8"] Nov 25 10:16:07 crc kubenswrapper[4926]: E1125 10:16:07.605499 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f34995-35ea-4875-9a69-61196081a80f" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.605525 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f34995-35ea-4875-9a69-61196081a80f" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.605770 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f34995-35ea-4875-9a69-61196081a80f" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.606575 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.612485 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8"] Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.614344 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.614487 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.614582 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.614745 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.699961 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.700079 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.700199 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pzrk\" (UniqueName: \"kubernetes.io/projected/0907de16-2f17-473e-83af-94baeb44725c-kube-api-access-6pzrk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.801775 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pzrk\" (UniqueName: \"kubernetes.io/projected/0907de16-2f17-473e-83af-94baeb44725c-kube-api-access-6pzrk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.801886 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.801953 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.807652 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.807647 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.822214 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pzrk\" (UniqueName: \"kubernetes.io/projected/0907de16-2f17-473e-83af-94baeb44725c-kube-api-access-6pzrk\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:07 crc kubenswrapper[4926]: I1125 10:16:07.925900 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:08 crc kubenswrapper[4926]: I1125 10:16:08.447034 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8"] Nov 25 10:16:08 crc kubenswrapper[4926]: I1125 10:16:08.552290 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" event={"ID":"0907de16-2f17-473e-83af-94baeb44725c","Type":"ContainerStarted","Data":"a9bf35d85c6250f8488023e772ccebd51dccb23abefd74fcb7a6a2ab178bdc0b"} Nov 25 10:16:09 crc kubenswrapper[4926]: I1125 10:16:09.562803 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" event={"ID":"0907de16-2f17-473e-83af-94baeb44725c","Type":"ContainerStarted","Data":"9ff9c9bfef636b1550905831efa21d568021a1bc1e7b3eea3599ea8d3e83ea2e"} Nov 25 10:16:09 crc kubenswrapper[4926]: I1125 10:16:09.580699 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" podStartSLOduration=2.146645162 podStartE2EDuration="2.580675979s" podCreationTimestamp="2025-11-25 10:16:07 +0000 UTC" firstStartedPulling="2025-11-25 10:16:08.453431979 +0000 UTC m=+1665.137388396" lastFinishedPulling="2025-11-25 10:16:08.887462786 +0000 UTC m=+1665.571419213" observedRunningTime="2025-11-25 10:16:09.580574527 +0000 UTC m=+1666.264530954" watchObservedRunningTime="2025-11-25 10:16:09.580675979 +0000 UTC m=+1666.264632406" Nov 25 10:16:15 crc kubenswrapper[4926]: I1125 10:16:15.012581 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:16:15 crc kubenswrapper[4926]: E1125 10:16:15.013312 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:16:22 crc kubenswrapper[4926]: I1125 10:16:22.056803 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-lslvn"] Nov 25 10:16:22 crc kubenswrapper[4926]: I1125 10:16:22.065447 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-lslvn"] Nov 25 10:16:24 crc kubenswrapper[4926]: I1125 10:16:24.021377 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb63e765-910d-4764-942c-bec556b98a5e" path="/var/lib/kubelet/pods/eb63e765-910d-4764-942c-bec556b98a5e/volumes" Nov 25 10:16:25 crc kubenswrapper[4926]: I1125 10:16:25.026617 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-cjmn2"] Nov 25 10:16:25 crc kubenswrapper[4926]: I1125 10:16:25.033874 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-cjmn2"] Nov 25 10:16:26 crc kubenswrapper[4926]: I1125 10:16:26.023155 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c73e4f7-21d9-4d05-b900-1e49498b4874" path="/var/lib/kubelet/pods/8c73e4f7-21d9-4d05-b900-1e49498b4874/volumes" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.166678 4926 scope.go:117] "RemoveContainer" containerID="e9533031e180f6e5b0ae28a161aec469b550ffce26995a2f7cdae035505f1001" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.221326 4926 scope.go:117] "RemoveContainer" containerID="39f7fd2c0764b76e4a6cd1fef98fdcb90f7f6372ffcb67355393cef922e722bc" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.257902 4926 scope.go:117] "RemoveContainer" containerID="35ec8108122ed4ff7547d52a2e40d1e3ea39c8053a2374ba5c5e23b909a87970" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.303011 4926 scope.go:117] "RemoveContainer" containerID="3cccb390990e940eaa77f32f2e691b926b51ce574babcfe23b92f4df1aa76c28" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.336773 4926 scope.go:117] "RemoveContainer" containerID="d9288af6d592a262ca6d7f702f7a6c4c0d258a53e7b71269db47ea91ecc56728" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.380495 4926 scope.go:117] "RemoveContainer" containerID="b092152b6402577975beddfefb125dea0f49ebaf85b3d481068bfcad3e2a5709" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.409242 4926 scope.go:117] "RemoveContainer" containerID="e6765602c9bbf756c103dde6d916c27b17acc611e585e6519f035172ff9e1063" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.442854 4926 scope.go:117] "RemoveContainer" containerID="b5415d070ca1c8776aa8d3d432c5e24a9220a9732426014a618911030b7c22cf" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.463536 4926 scope.go:117] "RemoveContainer" containerID="6b2c2d5a9644a0e7be4f78d1877d23621dda6d82b88e85a3d71338fbf868a486" Nov 25 10:16:27 crc kubenswrapper[4926]: I1125 10:16:27.486543 4926 scope.go:117] "RemoveContainer" containerID="2503491034c10378141b596e162960a67b1158b9e06d0a02b70f3a18377cb69d" Nov 25 10:16:28 crc kubenswrapper[4926]: I1125 10:16:28.012522 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:16:28 crc kubenswrapper[4926]: E1125 10:16:28.012838 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:16:31 crc kubenswrapper[4926]: I1125 10:16:31.032304 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-9fjls"] Nov 25 10:16:31 crc kubenswrapper[4926]: I1125 10:16:31.039320 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-9fjls"] Nov 25 10:16:32 crc kubenswrapper[4926]: I1125 10:16:32.022403 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="545481d3-3362-45f8-929b-71e8c75490dc" path="/var/lib/kubelet/pods/545481d3-3362-45f8-929b-71e8c75490dc/volumes" Nov 25 10:16:41 crc kubenswrapper[4926]: I1125 10:16:41.033631 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-k58gt"] Nov 25 10:16:41 crc kubenswrapper[4926]: I1125 10:16:41.042703 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-k58gt"] Nov 25 10:16:42 crc kubenswrapper[4926]: I1125 10:16:42.012597 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:16:42 crc kubenswrapper[4926]: E1125 10:16:42.012846 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:16:42 crc kubenswrapper[4926]: I1125 10:16:42.024811 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af3d71a6-8937-4d95-9152-d77c1afc39eb" path="/var/lib/kubelet/pods/af3d71a6-8937-4d95-9152-d77c1afc39eb/volumes" Nov 25 10:16:49 crc kubenswrapper[4926]: I1125 10:16:49.057089 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-k6rtr"] Nov 25 10:16:49 crc kubenswrapper[4926]: I1125 10:16:49.066457 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-k6rtr"] Nov 25 10:16:50 crc kubenswrapper[4926]: I1125 10:16:50.022701 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65e6c095-3a19-4536-b55a-c79c991badd0" path="/var/lib/kubelet/pods/65e6c095-3a19-4536-b55a-c79c991badd0/volumes" Nov 25 10:16:56 crc kubenswrapper[4926]: I1125 10:16:56.934582 4926 generic.go:334] "Generic (PLEG): container finished" podID="0907de16-2f17-473e-83af-94baeb44725c" containerID="9ff9c9bfef636b1550905831efa21d568021a1bc1e7b3eea3599ea8d3e83ea2e" exitCode=0 Nov 25 10:16:56 crc kubenswrapper[4926]: I1125 10:16:56.934699 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" event={"ID":"0907de16-2f17-473e-83af-94baeb44725c","Type":"ContainerDied","Data":"9ff9c9bfef636b1550905831efa21d568021a1bc1e7b3eea3599ea8d3e83ea2e"} Nov 25 10:16:57 crc kubenswrapper[4926]: I1125 10:16:57.012962 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:16:57 crc kubenswrapper[4926]: E1125 10:16:57.013253 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.309147 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.393382 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-ssh-key\") pod \"0907de16-2f17-473e-83af-94baeb44725c\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.393449 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-inventory\") pod \"0907de16-2f17-473e-83af-94baeb44725c\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.393473 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pzrk\" (UniqueName: \"kubernetes.io/projected/0907de16-2f17-473e-83af-94baeb44725c-kube-api-access-6pzrk\") pod \"0907de16-2f17-473e-83af-94baeb44725c\" (UID: \"0907de16-2f17-473e-83af-94baeb44725c\") " Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.399048 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0907de16-2f17-473e-83af-94baeb44725c-kube-api-access-6pzrk" (OuterVolumeSpecName: "kube-api-access-6pzrk") pod "0907de16-2f17-473e-83af-94baeb44725c" (UID: "0907de16-2f17-473e-83af-94baeb44725c"). InnerVolumeSpecName "kube-api-access-6pzrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.420055 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-inventory" (OuterVolumeSpecName: "inventory") pod "0907de16-2f17-473e-83af-94baeb44725c" (UID: "0907de16-2f17-473e-83af-94baeb44725c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.433274 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0907de16-2f17-473e-83af-94baeb44725c" (UID: "0907de16-2f17-473e-83af-94baeb44725c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.496735 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.496769 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0907de16-2f17-473e-83af-94baeb44725c-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.496795 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pzrk\" (UniqueName: \"kubernetes.io/projected/0907de16-2f17-473e-83af-94baeb44725c-kube-api-access-6pzrk\") on node \"crc\" DevicePath \"\"" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.985785 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" event={"ID":"0907de16-2f17-473e-83af-94baeb44725c","Type":"ContainerDied","Data":"a9bf35d85c6250f8488023e772ccebd51dccb23abefd74fcb7a6a2ab178bdc0b"} Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.985826 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9bf35d85c6250f8488023e772ccebd51dccb23abefd74fcb7a6a2ab178bdc0b" Nov 25 10:16:58 crc kubenswrapper[4926]: I1125 10:16:58.985888 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.040707 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5fxd9"] Nov 25 10:16:59 crc kubenswrapper[4926]: E1125 10:16:59.041145 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0907de16-2f17-473e-83af-94baeb44725c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.041164 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="0907de16-2f17-473e-83af-94baeb44725c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.041487 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="0907de16-2f17-473e-83af-94baeb44725c" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.042085 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.045723 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.045768 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.045895 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.046029 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.049582 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5fxd9"] Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.107037 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnp52\" (UniqueName: \"kubernetes.io/projected/6f0df4f6-baeb-4c84-baa9-551970e17317-kube-api-access-bnp52\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.107190 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.107227 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.208177 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnp52\" (UniqueName: \"kubernetes.io/projected/6f0df4f6-baeb-4c84-baa9-551970e17317-kube-api-access-bnp52\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.208246 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.208269 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.212428 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.215083 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.243429 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnp52\" (UniqueName: \"kubernetes.io/projected/6f0df4f6-baeb-4c84-baa9-551970e17317-kube-api-access-bnp52\") pod \"ssh-known-hosts-edpm-deployment-5fxd9\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.364749 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.876622 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-5fxd9"] Nov 25 10:16:59 crc kubenswrapper[4926]: I1125 10:16:59.997366 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" event={"ID":"6f0df4f6-baeb-4c84-baa9-551970e17317","Type":"ContainerStarted","Data":"ba8fa85c5f272914b83353644710d1fea4bcbb65a9323878d1589731f74e4f07"} Nov 25 10:17:02 crc kubenswrapper[4926]: I1125 10:17:02.022266 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" event={"ID":"6f0df4f6-baeb-4c84-baa9-551970e17317","Type":"ContainerStarted","Data":"c02635af16f38e0f1808cbbbec5581e3f000a382c55db0f5f4cbea0d4366d7ff"} Nov 25 10:17:08 crc kubenswrapper[4926]: I1125 10:17:08.061618 4926 generic.go:334] "Generic (PLEG): container finished" podID="6f0df4f6-baeb-4c84-baa9-551970e17317" containerID="c02635af16f38e0f1808cbbbec5581e3f000a382c55db0f5f4cbea0d4366d7ff" exitCode=0 Nov 25 10:17:08 crc kubenswrapper[4926]: I1125 10:17:08.061655 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" event={"ID":"6f0df4f6-baeb-4c84-baa9-551970e17317","Type":"ContainerDied","Data":"c02635af16f38e0f1808cbbbec5581e3f000a382c55db0f5f4cbea0d4366d7ff"} Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.894801 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.896785 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-ssh-key-openstack-edpm-ipam\") pod \"6f0df4f6-baeb-4c84-baa9-551970e17317\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.896889 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnp52\" (UniqueName: \"kubernetes.io/projected/6f0df4f6-baeb-4c84-baa9-551970e17317-kube-api-access-bnp52\") pod \"6f0df4f6-baeb-4c84-baa9-551970e17317\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.896989 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-inventory-0\") pod \"6f0df4f6-baeb-4c84-baa9-551970e17317\" (UID: \"6f0df4f6-baeb-4c84-baa9-551970e17317\") " Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.902241 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f0df4f6-baeb-4c84-baa9-551970e17317-kube-api-access-bnp52" (OuterVolumeSpecName: "kube-api-access-bnp52") pod "6f0df4f6-baeb-4c84-baa9-551970e17317" (UID: "6f0df4f6-baeb-4c84-baa9-551970e17317"). InnerVolumeSpecName "kube-api-access-bnp52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.923708 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "6f0df4f6-baeb-4c84-baa9-551970e17317" (UID: "6f0df4f6-baeb-4c84-baa9-551970e17317"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.931833 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "6f0df4f6-baeb-4c84-baa9-551970e17317" (UID: "6f0df4f6-baeb-4c84-baa9-551970e17317"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.998634 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.998707 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnp52\" (UniqueName: \"kubernetes.io/projected/6f0df4f6-baeb-4c84-baa9-551970e17317-kube-api-access-bnp52\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:09 crc kubenswrapper[4926]: I1125 10:17:09.998719 4926 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/6f0df4f6-baeb-4c84-baa9-551970e17317-inventory-0\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.077134 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" event={"ID":"6f0df4f6-baeb-4c84-baa9-551970e17317","Type":"ContainerDied","Data":"ba8fa85c5f272914b83353644710d1fea4bcbb65a9323878d1589731f74e4f07"} Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.077184 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba8fa85c5f272914b83353644710d1fea4bcbb65a9323878d1589731f74e4f07" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.077186 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-5fxd9" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.154097 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7"] Nov 25 10:17:10 crc kubenswrapper[4926]: E1125 10:17:10.154485 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f0df4f6-baeb-4c84-baa9-551970e17317" containerName="ssh-known-hosts-edpm-deployment" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.154504 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f0df4f6-baeb-4c84-baa9-551970e17317" containerName="ssh-known-hosts-edpm-deployment" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.154689 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f0df4f6-baeb-4c84-baa9-551970e17317" containerName="ssh-known-hosts-edpm-deployment" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.155300 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.160576 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.160633 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.160863 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.161033 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.164405 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7"] Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.307141 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.307219 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xwjh\" (UniqueName: \"kubernetes.io/projected/f04c96be-4b0e-43a9-9362-90c38248ac72-kube-api-access-2xwjh\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.307265 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.408750 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.408791 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xwjh\" (UniqueName: \"kubernetes.io/projected/f04c96be-4b0e-43a9-9362-90c38248ac72-kube-api-access-2xwjh\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.408809 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.413377 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.413531 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.425845 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xwjh\" (UniqueName: \"kubernetes.io/projected/f04c96be-4b0e-43a9-9362-90c38248ac72-kube-api-access-2xwjh\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-dt7h7\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.479751 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:10 crc kubenswrapper[4926]: I1125 10:17:10.954253 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7"] Nov 25 10:17:11 crc kubenswrapper[4926]: I1125 10:17:11.012252 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:17:11 crc kubenswrapper[4926]: E1125 10:17:11.012567 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:17:11 crc kubenswrapper[4926]: I1125 10:17:11.086503 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" event={"ID":"f04c96be-4b0e-43a9-9362-90c38248ac72","Type":"ContainerStarted","Data":"32a34fd17088f84c31b7e015c61287aaecf24436723d26a2375a57766e492368"} Nov 25 10:17:12 crc kubenswrapper[4926]: I1125 10:17:12.040808 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pr557"] Nov 25 10:17:12 crc kubenswrapper[4926]: I1125 10:17:12.048553 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pr557"] Nov 25 10:17:12 crc kubenswrapper[4926]: I1125 10:17:12.095849 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" event={"ID":"f04c96be-4b0e-43a9-9362-90c38248ac72","Type":"ContainerStarted","Data":"a485916b59b7a8b9932b2ae0779d4e62dfd2ea3ad369ae7bc25cdeb4010c0f28"} Nov 25 10:17:12 crc kubenswrapper[4926]: I1125 10:17:12.113210 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" podStartSLOduration=1.64281282 podStartE2EDuration="2.11319237s" podCreationTimestamp="2025-11-25 10:17:10 +0000 UTC" firstStartedPulling="2025-11-25 10:17:10.960495648 +0000 UTC m=+1727.644452075" lastFinishedPulling="2025-11-25 10:17:11.430875198 +0000 UTC m=+1728.114831625" observedRunningTime="2025-11-25 10:17:12.108433502 +0000 UTC m=+1728.792389929" watchObservedRunningTime="2025-11-25 10:17:12.11319237 +0000 UTC m=+1728.797148797" Nov 25 10:17:14 crc kubenswrapper[4926]: I1125 10:17:14.021835 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc" path="/var/lib/kubelet/pods/1ec1ac8d-a9de-48eb-ba61-a0cfccaf46bc/volumes" Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.026820 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-dded-account-create-scxpg"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.033573 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f822-account-create-hghgh"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.048135 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-w4qxc"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.055052 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0b02-account-create-jpmcd"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.062461 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-dded-account-create-scxpg"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.069129 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f822-account-create-hghgh"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.074883 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-jqn4l"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.082341 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0b02-account-create-jpmcd"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.088698 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-w4qxc"] Nov 25 10:17:18 crc kubenswrapper[4926]: I1125 10:17:18.095111 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-jqn4l"] Nov 25 10:17:19 crc kubenswrapper[4926]: I1125 10:17:19.152290 4926 generic.go:334] "Generic (PLEG): container finished" podID="f04c96be-4b0e-43a9-9362-90c38248ac72" containerID="a485916b59b7a8b9932b2ae0779d4e62dfd2ea3ad369ae7bc25cdeb4010c0f28" exitCode=0 Nov 25 10:17:19 crc kubenswrapper[4926]: I1125 10:17:19.152400 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" event={"ID":"f04c96be-4b0e-43a9-9362-90c38248ac72","Type":"ContainerDied","Data":"a485916b59b7a8b9932b2ae0779d4e62dfd2ea3ad369ae7bc25cdeb4010c0f28"} Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.031342 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e7b372d-b0e0-4423-9734-00d223b0e99e" path="/var/lib/kubelet/pods/2e7b372d-b0e0-4423-9734-00d223b0e99e/volumes" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.032413 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81121ebf-1d21-43de-8787-efc9b79578e2" path="/var/lib/kubelet/pods/81121ebf-1d21-43de-8787-efc9b79578e2/volumes" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.033372 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e32ccd7-0d61-4328-a22b-4637bd3f16f4" path="/var/lib/kubelet/pods/8e32ccd7-0d61-4328-a22b-4637bd3f16f4/volumes" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.034352 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c58cf150-5525-4d1c-9376-a3bd78100269" path="/var/lib/kubelet/pods/c58cf150-5525-4d1c-9376-a3bd78100269/volumes" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.035864 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7d24462-3c87-4454-bf92-e97345e32e12" path="/var/lib/kubelet/pods/c7d24462-3c87-4454-bf92-e97345e32e12/volumes" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.531240 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.698102 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-ssh-key\") pod \"f04c96be-4b0e-43a9-9362-90c38248ac72\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.698266 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-inventory\") pod \"f04c96be-4b0e-43a9-9362-90c38248ac72\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.698304 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xwjh\" (UniqueName: \"kubernetes.io/projected/f04c96be-4b0e-43a9-9362-90c38248ac72-kube-api-access-2xwjh\") pod \"f04c96be-4b0e-43a9-9362-90c38248ac72\" (UID: \"f04c96be-4b0e-43a9-9362-90c38248ac72\") " Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.704210 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f04c96be-4b0e-43a9-9362-90c38248ac72-kube-api-access-2xwjh" (OuterVolumeSpecName: "kube-api-access-2xwjh") pod "f04c96be-4b0e-43a9-9362-90c38248ac72" (UID: "f04c96be-4b0e-43a9-9362-90c38248ac72"). InnerVolumeSpecName "kube-api-access-2xwjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.730150 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-inventory" (OuterVolumeSpecName: "inventory") pod "f04c96be-4b0e-43a9-9362-90c38248ac72" (UID: "f04c96be-4b0e-43a9-9362-90c38248ac72"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.730451 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f04c96be-4b0e-43a9-9362-90c38248ac72" (UID: "f04c96be-4b0e-43a9-9362-90c38248ac72"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.801071 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.801117 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xwjh\" (UniqueName: \"kubernetes.io/projected/f04c96be-4b0e-43a9-9362-90c38248ac72-kube-api-access-2xwjh\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:20 crc kubenswrapper[4926]: I1125 10:17:20.801169 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f04c96be-4b0e-43a9-9362-90c38248ac72-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.172757 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" event={"ID":"f04c96be-4b0e-43a9-9362-90c38248ac72","Type":"ContainerDied","Data":"32a34fd17088f84c31b7e015c61287aaecf24436723d26a2375a57766e492368"} Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.172802 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32a34fd17088f84c31b7e015c61287aaecf24436723d26a2375a57766e492368" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.172906 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-dt7h7" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.229018 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t"] Nov 25 10:17:21 crc kubenswrapper[4926]: E1125 10:17:21.229469 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f04c96be-4b0e-43a9-9362-90c38248ac72" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.229495 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f04c96be-4b0e-43a9-9362-90c38248ac72" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.229719 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f04c96be-4b0e-43a9-9362-90c38248ac72" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.230449 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.235710 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.235921 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-knp2x" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.236609 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.237100 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.238435 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t"] Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.411128 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.411266 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c4qk\" (UniqueName: \"kubernetes.io/projected/c2514ec8-a2ab-485d-9385-838ac88416d0-kube-api-access-9c4qk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.411313 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.512221 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.512323 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c4qk\" (UniqueName: \"kubernetes.io/projected/c2514ec8-a2ab-485d-9385-838ac88416d0-kube-api-access-9c4qk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.512367 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.515768 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.516415 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.532580 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c4qk\" (UniqueName: \"kubernetes.io/projected/c2514ec8-a2ab-485d-9385-838ac88416d0-kube-api-access-9c4qk\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:21 crc kubenswrapper[4926]: I1125 10:17:21.579816 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:22 crc kubenswrapper[4926]: I1125 10:17:22.080352 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t"] Nov 25 10:17:22 crc kubenswrapper[4926]: I1125 10:17:22.180725 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" event={"ID":"c2514ec8-a2ab-485d-9385-838ac88416d0","Type":"ContainerStarted","Data":"c189fded8cf2f3200342d424104224c9ff6789ec6eb619e5f3d81b9854f614f5"} Nov 25 10:17:23 crc kubenswrapper[4926]: I1125 10:17:23.189950 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" event={"ID":"c2514ec8-a2ab-485d-9385-838ac88416d0","Type":"ContainerStarted","Data":"b25d9bbdc3197382e59069586af421431b479f83ffbe4dbd6a7430ccd0923f23"} Nov 25 10:17:25 crc kubenswrapper[4926]: I1125 10:17:25.012184 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:17:25 crc kubenswrapper[4926]: E1125 10:17:25.012709 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.742235 4926 scope.go:117] "RemoveContainer" containerID="96b38b7ef994115e7ebb8e7cf6948d5b45a0faee4a0cdb3c89f526104b81337d" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.783422 4926 scope.go:117] "RemoveContainer" containerID="2b44fa0a33b70b579973a762a3b624426c4acdea5da29fc400116e50d1330c67" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.805810 4926 scope.go:117] "RemoveContainer" containerID="893aa06815f5c11ee726bd49637d5a109d184e83b517cc17a67ea83e5e444eb5" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.841248 4926 scope.go:117] "RemoveContainer" containerID="e01be87fd1511b3fab3ca6d93a3ddcc5e8dee7080df81c037dfd317c9d2e6750" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.878037 4926 scope.go:117] "RemoveContainer" containerID="8f1ac120062adb0fbd34758fc7c784ee691a7138d7e707fcb64648157b587bd0" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.926886 4926 scope.go:117] "RemoveContainer" containerID="5c92d628d09875a45b90e71d2da83c4d2f7da2dfed7704d011f9786db175b20c" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.953167 4926 scope.go:117] "RemoveContainer" containerID="75e798cfda64009187e2d64cf60d99cae626e1b9a49900c8cb226074133c8fde" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.972474 4926 scope.go:117] "RemoveContainer" containerID="392003ba1866bcf1d975a622e4aae4a9d6bbe1ccb80a0561fbee4f4e281f25b4" Nov 25 10:17:27 crc kubenswrapper[4926]: I1125 10:17:27.996995 4926 scope.go:117] "RemoveContainer" containerID="35db8bbb0f786e7c87bad0f7902281aed6b6f6ab4b9e41939aac0ac33819b745" Nov 25 10:17:32 crc kubenswrapper[4926]: I1125 10:17:32.263719 4926 generic.go:334] "Generic (PLEG): container finished" podID="c2514ec8-a2ab-485d-9385-838ac88416d0" containerID="b25d9bbdc3197382e59069586af421431b479f83ffbe4dbd6a7430ccd0923f23" exitCode=0 Nov 25 10:17:32 crc kubenswrapper[4926]: I1125 10:17:32.263798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" event={"ID":"c2514ec8-a2ab-485d-9385-838ac88416d0","Type":"ContainerDied","Data":"b25d9bbdc3197382e59069586af421431b479f83ffbe4dbd6a7430ccd0923f23"} Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.649918 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.828311 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-ssh-key\") pod \"c2514ec8-a2ab-485d-9385-838ac88416d0\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.828407 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c4qk\" (UniqueName: \"kubernetes.io/projected/c2514ec8-a2ab-485d-9385-838ac88416d0-kube-api-access-9c4qk\") pod \"c2514ec8-a2ab-485d-9385-838ac88416d0\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.828493 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-inventory\") pod \"c2514ec8-a2ab-485d-9385-838ac88416d0\" (UID: \"c2514ec8-a2ab-485d-9385-838ac88416d0\") " Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.836911 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2514ec8-a2ab-485d-9385-838ac88416d0-kube-api-access-9c4qk" (OuterVolumeSpecName: "kube-api-access-9c4qk") pod "c2514ec8-a2ab-485d-9385-838ac88416d0" (UID: "c2514ec8-a2ab-485d-9385-838ac88416d0"). InnerVolumeSpecName "kube-api-access-9c4qk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.854174 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c2514ec8-a2ab-485d-9385-838ac88416d0" (UID: "c2514ec8-a2ab-485d-9385-838ac88416d0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.854548 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-inventory" (OuterVolumeSpecName: "inventory") pod "c2514ec8-a2ab-485d-9385-838ac88416d0" (UID: "c2514ec8-a2ab-485d-9385-838ac88416d0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.930155 4926 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-ssh-key\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.930203 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c4qk\" (UniqueName: \"kubernetes.io/projected/c2514ec8-a2ab-485d-9385-838ac88416d0-kube-api-access-9c4qk\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:33 crc kubenswrapper[4926]: I1125 10:17:33.930222 4926 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c2514ec8-a2ab-485d-9385-838ac88416d0-inventory\") on node \"crc\" DevicePath \"\"" Nov 25 10:17:34 crc kubenswrapper[4926]: I1125 10:17:34.284120 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" event={"ID":"c2514ec8-a2ab-485d-9385-838ac88416d0","Type":"ContainerDied","Data":"c189fded8cf2f3200342d424104224c9ff6789ec6eb619e5f3d81b9854f614f5"} Nov 25 10:17:34 crc kubenswrapper[4926]: I1125 10:17:34.284166 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c189fded8cf2f3200342d424104224c9ff6789ec6eb619e5f3d81b9854f614f5" Nov 25 10:17:34 crc kubenswrapper[4926]: I1125 10:17:34.284166 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t" Nov 25 10:17:39 crc kubenswrapper[4926]: I1125 10:17:39.012617 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:17:39 crc kubenswrapper[4926]: E1125 10:17:39.013504 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:17:48 crc kubenswrapper[4926]: I1125 10:17:48.043989 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-twczw"] Nov 25 10:17:48 crc kubenswrapper[4926]: I1125 10:17:48.053178 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-twczw"] Nov 25 10:17:50 crc kubenswrapper[4926]: I1125 10:17:50.013050 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:17:50 crc kubenswrapper[4926]: E1125 10:17:50.013540 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:17:50 crc kubenswrapper[4926]: I1125 10:17:50.022774 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a33ef41d-bc58-413a-946a-41e0a85450e5" path="/var/lib/kubelet/pods/a33ef41d-bc58-413a-946a-41e0a85450e5/volumes" Nov 25 10:18:05 crc kubenswrapper[4926]: I1125 10:18:05.013335 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:18:05 crc kubenswrapper[4926]: E1125 10:18:05.014063 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:18:11 crc kubenswrapper[4926]: I1125 10:18:11.028054 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lb4mw"] Nov 25 10:18:11 crc kubenswrapper[4926]: I1125 10:18:11.034880 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-xqpfc"] Nov 25 10:18:11 crc kubenswrapper[4926]: I1125 10:18:11.042417 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-lb4mw"] Nov 25 10:18:11 crc kubenswrapper[4926]: I1125 10:18:11.048803 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-xqpfc"] Nov 25 10:18:12 crc kubenswrapper[4926]: I1125 10:18:12.022065 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29ce8c5b-26dd-47bc-b6a4-19e99a313859" path="/var/lib/kubelet/pods/29ce8c5b-26dd-47bc-b6a4-19e99a313859/volumes" Nov 25 10:18:12 crc kubenswrapper[4926]: I1125 10:18:12.022700 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c43314e-7f71-40cb-ab8d-8ac5168e5622" path="/var/lib/kubelet/pods/4c43314e-7f71-40cb-ab8d-8ac5168e5622/volumes" Nov 25 10:18:19 crc kubenswrapper[4926]: I1125 10:18:19.022862 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:18:19 crc kubenswrapper[4926]: E1125 10:18:19.024058 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.116098 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vw4vz"] Nov 25 10:18:27 crc kubenswrapper[4926]: E1125 10:18:27.117553 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2514ec8-a2ab-485d-9385-838ac88416d0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.117576 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2514ec8-a2ab-485d-9385-838ac88416d0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.117965 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2514ec8-a2ab-485d-9385-838ac88416d0" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.119732 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.138549 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vw4vz"] Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.172669 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.172740 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2lqw\" (UniqueName: \"kubernetes.io/projected/05218437-b051-4be9-9a59-89a44ef50530-kube-api-access-z2lqw\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.172868 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-utilities\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.274786 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-utilities\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.274859 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.274916 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2lqw\" (UniqueName: \"kubernetes.io/projected/05218437-b051-4be9-9a59-89a44ef50530-kube-api-access-z2lqw\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.275368 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-utilities\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.275421 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.294554 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2lqw\" (UniqueName: \"kubernetes.io/projected/05218437-b051-4be9-9a59-89a44ef50530-kube-api-access-z2lqw\") pod \"redhat-operators-vw4vz\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.311396 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-xxnjf"] Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.313475 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.324889 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxnjf"] Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.375626 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-utilities\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.375699 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4kkk\" (UniqueName: \"kubernetes.io/projected/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-kube-api-access-v4kkk\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.375778 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-catalog-content\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.477595 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-utilities\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.477689 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4kkk\" (UniqueName: \"kubernetes.io/projected/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-kube-api-access-v4kkk\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.477743 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-catalog-content\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.478099 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-utilities\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.478194 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-catalog-content\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.490808 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.495581 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4kkk\" (UniqueName: \"kubernetes.io/projected/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-kube-api-access-v4kkk\") pod \"certified-operators-xxnjf\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.659316 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:27 crc kubenswrapper[4926]: I1125 10:18:27.963282 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vw4vz"] Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.146700 4926 scope.go:117] "RemoveContainer" containerID="9d5ac90b538ee9d752eedfe0425ee54eaff126c84a7dbaae8e2042a050f42521" Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.209927 4926 scope.go:117] "RemoveContainer" containerID="41b013eed5424ea88277aed24dc581155f1ce52b8c30ca28c8d549b346b4f4ef" Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.213830 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-xxnjf"] Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.270636 4926 scope.go:117] "RemoveContainer" containerID="67010eccfa60f625268c622a8ac0a3aa8e11f411f5da4021ea963988942a6ca7" Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.807268 4926 generic.go:334] "Generic (PLEG): container finished" podID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerID="e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212" exitCode=0 Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.807361 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerDied","Data":"e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212"} Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.807628 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerStarted","Data":"9ef7cbed24aaa5a4d6c62fae20267f0188d1803b5238395a85883e00a9c70cdb"} Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.810026 4926 generic.go:334] "Generic (PLEG): container finished" podID="05218437-b051-4be9-9a59-89a44ef50530" containerID="2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3" exitCode=0 Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.810079 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw4vz" event={"ID":"05218437-b051-4be9-9a59-89a44ef50530","Type":"ContainerDied","Data":"2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3"} Nov 25 10:18:28 crc kubenswrapper[4926]: I1125 10:18:28.810108 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw4vz" event={"ID":"05218437-b051-4be9-9a59-89a44ef50530","Type":"ContainerStarted","Data":"c385d221ec1a5bcd33ff5ee966b4c772c5869f62a4699bbef624056f058c529a"} Nov 25 10:18:29 crc kubenswrapper[4926]: I1125 10:18:29.819044 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerStarted","Data":"2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2"} Nov 25 10:18:30 crc kubenswrapper[4926]: I1125 10:18:30.828330 4926 generic.go:334] "Generic (PLEG): container finished" podID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerID="2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2" exitCode=0 Nov 25 10:18:30 crc kubenswrapper[4926]: I1125 10:18:30.828428 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerDied","Data":"2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2"} Nov 25 10:18:30 crc kubenswrapper[4926]: I1125 10:18:30.848092 4926 generic.go:334] "Generic (PLEG): container finished" podID="05218437-b051-4be9-9a59-89a44ef50530" containerID="f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2" exitCode=0 Nov 25 10:18:30 crc kubenswrapper[4926]: I1125 10:18:30.848136 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw4vz" event={"ID":"05218437-b051-4be9-9a59-89a44ef50530","Type":"ContainerDied","Data":"f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2"} Nov 25 10:18:31 crc kubenswrapper[4926]: I1125 10:18:31.859262 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerStarted","Data":"7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8"} Nov 25 10:18:31 crc kubenswrapper[4926]: I1125 10:18:31.861698 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw4vz" event={"ID":"05218437-b051-4be9-9a59-89a44ef50530","Type":"ContainerStarted","Data":"f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931"} Nov 25 10:18:31 crc kubenswrapper[4926]: I1125 10:18:31.884773 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-xxnjf" podStartSLOduration=2.323429789 podStartE2EDuration="4.884751338s" podCreationTimestamp="2025-11-25 10:18:27 +0000 UTC" firstStartedPulling="2025-11-25 10:18:28.809854837 +0000 UTC m=+1805.493811254" lastFinishedPulling="2025-11-25 10:18:31.371176376 +0000 UTC m=+1808.055132803" observedRunningTime="2025-11-25 10:18:31.883130618 +0000 UTC m=+1808.567087045" watchObservedRunningTime="2025-11-25 10:18:31.884751338 +0000 UTC m=+1808.568707785" Nov 25 10:18:31 crc kubenswrapper[4926]: I1125 10:18:31.910595 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vw4vz" podStartSLOduration=2.397849286 podStartE2EDuration="4.910574239s" podCreationTimestamp="2025-11-25 10:18:27 +0000 UTC" firstStartedPulling="2025-11-25 10:18:28.812123923 +0000 UTC m=+1805.496080350" lastFinishedPulling="2025-11-25 10:18:31.324848876 +0000 UTC m=+1808.008805303" observedRunningTime="2025-11-25 10:18:31.905515053 +0000 UTC m=+1808.589471490" watchObservedRunningTime="2025-11-25 10:18:31.910574239 +0000 UTC m=+1808.594530666" Nov 25 10:18:34 crc kubenswrapper[4926]: I1125 10:18:34.016990 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:18:34 crc kubenswrapper[4926]: E1125 10:18:34.017537 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.491689 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.492331 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.534494 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.660665 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.660714 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.713299 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.957039 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:37 crc kubenswrapper[4926]: I1125 10:18:37.957730 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:39 crc kubenswrapper[4926]: I1125 10:18:39.384831 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxnjf"] Nov 25 10:18:39 crc kubenswrapper[4926]: I1125 10:18:39.924956 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-xxnjf" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="registry-server" containerID="cri-o://7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8" gracePeriod=2 Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.369233 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vw4vz"] Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.427026 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.608794 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4kkk\" (UniqueName: \"kubernetes.io/projected/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-kube-api-access-v4kkk\") pod \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.608996 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-catalog-content\") pod \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.609110 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-utilities\") pod \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\" (UID: \"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9\") " Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.610208 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-utilities" (OuterVolumeSpecName: "utilities") pod "8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" (UID: "8fb1b1de-dfd5-4afb-b31b-45e396de9dc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.619102 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-kube-api-access-v4kkk" (OuterVolumeSpecName: "kube-api-access-v4kkk") pod "8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" (UID: "8fb1b1de-dfd5-4afb-b31b-45e396de9dc9"). InnerVolumeSpecName "kube-api-access-v4kkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.663587 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" (UID: "8fb1b1de-dfd5-4afb-b31b-45e396de9dc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.711047 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.711088 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4kkk\" (UniqueName: \"kubernetes.io/projected/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-kube-api-access-v4kkk\") on node \"crc\" DevicePath \"\"" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.711101 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.942201 4926 generic.go:334] "Generic (PLEG): container finished" podID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerID="7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8" exitCode=0 Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.942411 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vw4vz" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="registry-server" containerID="cri-o://f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931" gracePeriod=2 Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.942677 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-xxnjf" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.944206 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerDied","Data":"7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8"} Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.944240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-xxnjf" event={"ID":"8fb1b1de-dfd5-4afb-b31b-45e396de9dc9","Type":"ContainerDied","Data":"9ef7cbed24aaa5a4d6c62fae20267f0188d1803b5238395a85883e00a9c70cdb"} Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.944256 4926 scope.go:117] "RemoveContainer" containerID="7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.975682 4926 scope.go:117] "RemoveContainer" containerID="2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2" Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.983011 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-xxnjf"] Nov 25 10:18:40 crc kubenswrapper[4926]: I1125 10:18:40.990099 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-xxnjf"] Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.017022 4926 scope.go:117] "RemoveContainer" containerID="e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.040390 4926 scope.go:117] "RemoveContainer" containerID="7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8" Nov 25 10:18:41 crc kubenswrapper[4926]: E1125 10:18:41.040780 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8\": container with ID starting with 7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8 not found: ID does not exist" containerID="7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.040806 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8"} err="failed to get container status \"7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8\": rpc error: code = NotFound desc = could not find container \"7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8\": container with ID starting with 7e927539ed8822dffa49f90507eef2a67fe72e48ec70f34364f3490d350348f8 not found: ID does not exist" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.040826 4926 scope.go:117] "RemoveContainer" containerID="2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2" Nov 25 10:18:41 crc kubenswrapper[4926]: E1125 10:18:41.041230 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2\": container with ID starting with 2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2 not found: ID does not exist" containerID="2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.041258 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2"} err="failed to get container status \"2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2\": rpc error: code = NotFound desc = could not find container \"2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2\": container with ID starting with 2a637520d7fa77dbe7705cdd3b52882e743d34a0eaaa93da534fcf324f264cb2 not found: ID does not exist" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.041275 4926 scope.go:117] "RemoveContainer" containerID="e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212" Nov 25 10:18:41 crc kubenswrapper[4926]: E1125 10:18:41.042563 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212\": container with ID starting with e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212 not found: ID does not exist" containerID="e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.042811 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212"} err="failed to get container status \"e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212\": rpc error: code = NotFound desc = could not find container \"e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212\": container with ID starting with e16e3d14192db270ddd05331f1f163e482e4cb2b2421c6b839d8497ed959c212 not found: ID does not exist" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.941039 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.957724 4926 generic.go:334] "Generic (PLEG): container finished" podID="05218437-b051-4be9-9a59-89a44ef50530" containerID="f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931" exitCode=0 Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.957767 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw4vz" event={"ID":"05218437-b051-4be9-9a59-89a44ef50530","Type":"ContainerDied","Data":"f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931"} Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.957798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw4vz" event={"ID":"05218437-b051-4be9-9a59-89a44ef50530","Type":"ContainerDied","Data":"c385d221ec1a5bcd33ff5ee966b4c772c5869f62a4699bbef624056f058c529a"} Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.957818 4926 scope.go:117] "RemoveContainer" containerID="f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.957971 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vw4vz" Nov 25 10:18:41 crc kubenswrapper[4926]: I1125 10:18:41.983043 4926 scope.go:117] "RemoveContainer" containerID="f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.024554 4926 scope.go:117] "RemoveContainer" containerID="2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.042196 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" path="/var/lib/kubelet/pods/8fb1b1de-dfd5-4afb-b31b-45e396de9dc9/volumes" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.060403 4926 scope.go:117] "RemoveContainer" containerID="f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931" Nov 25 10:18:42 crc kubenswrapper[4926]: E1125 10:18:42.060899 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931\": container with ID starting with f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931 not found: ID does not exist" containerID="f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.060969 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931"} err="failed to get container status \"f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931\": rpc error: code = NotFound desc = could not find container \"f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931\": container with ID starting with f9a541f8946ed71e66dcc38643247b0dd692497c8e473863f369d932e5e77931 not found: ID does not exist" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.060998 4926 scope.go:117] "RemoveContainer" containerID="f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2" Nov 25 10:18:42 crc kubenswrapper[4926]: E1125 10:18:42.061635 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2\": container with ID starting with f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2 not found: ID does not exist" containerID="f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.061664 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2"} err="failed to get container status \"f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2\": rpc error: code = NotFound desc = could not find container \"f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2\": container with ID starting with f3ceab230b0d756e70c0fbbfe766132d0f8809b42a27f41e6a89e316c41b67c2 not found: ID does not exist" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.061686 4926 scope.go:117] "RemoveContainer" containerID="2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3" Nov 25 10:18:42 crc kubenswrapper[4926]: E1125 10:18:42.062066 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3\": container with ID starting with 2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3 not found: ID does not exist" containerID="2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.062149 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3"} err="failed to get container status \"2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3\": rpc error: code = NotFound desc = could not find container \"2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3\": container with ID starting with 2822002a4a898658d16d194fd9fa1a581e0a1d24fe4f7511a1bf9fd5636798d3 not found: ID does not exist" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.146239 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content\") pod \"05218437-b051-4be9-9a59-89a44ef50530\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.146313 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2lqw\" (UniqueName: \"kubernetes.io/projected/05218437-b051-4be9-9a59-89a44ef50530-kube-api-access-z2lqw\") pod \"05218437-b051-4be9-9a59-89a44ef50530\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.146351 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-utilities\") pod \"05218437-b051-4be9-9a59-89a44ef50530\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.147774 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-utilities" (OuterVolumeSpecName: "utilities") pod "05218437-b051-4be9-9a59-89a44ef50530" (UID: "05218437-b051-4be9-9a59-89a44ef50530"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.152040 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05218437-b051-4be9-9a59-89a44ef50530-kube-api-access-z2lqw" (OuterVolumeSpecName: "kube-api-access-z2lqw") pod "05218437-b051-4be9-9a59-89a44ef50530" (UID: "05218437-b051-4be9-9a59-89a44ef50530"). InnerVolumeSpecName "kube-api-access-z2lqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.246969 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05218437-b051-4be9-9a59-89a44ef50530" (UID: "05218437-b051-4be9-9a59-89a44ef50530"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.247596 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content\") pod \"05218437-b051-4be9-9a59-89a44ef50530\" (UID: \"05218437-b051-4be9-9a59-89a44ef50530\") " Nov 25 10:18:42 crc kubenswrapper[4926]: W1125 10:18:42.247797 4926 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/05218437-b051-4be9-9a59-89a44ef50530/volumes/kubernetes.io~empty-dir/catalog-content Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.247832 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "05218437-b051-4be9-9a59-89a44ef50530" (UID: "05218437-b051-4be9-9a59-89a44ef50530"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.248149 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2lqw\" (UniqueName: \"kubernetes.io/projected/05218437-b051-4be9-9a59-89a44ef50530-kube-api-access-z2lqw\") on node \"crc\" DevicePath \"\"" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.248167 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.248177 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/05218437-b051-4be9-9a59-89a44ef50530-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.290386 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vw4vz"] Nov 25 10:18:42 crc kubenswrapper[4926]: I1125 10:18:42.296040 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vw4vz"] Nov 25 10:18:44 crc kubenswrapper[4926]: I1125 10:18:44.022786 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05218437-b051-4be9-9a59-89a44ef50530" path="/var/lib/kubelet/pods/05218437-b051-4be9-9a59-89a44ef50530/volumes" Nov 25 10:18:48 crc kubenswrapper[4926]: I1125 10:18:48.012839 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:18:49 crc kubenswrapper[4926]: I1125 10:18:49.017138 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"d6f381f071189398b29183f9732862d78affd0630860a5419ed3e39c642baae0"} Nov 25 10:18:55 crc kubenswrapper[4926]: I1125 10:18:55.055776 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-mzsx6"] Nov 25 10:18:55 crc kubenswrapper[4926]: I1125 10:18:55.066993 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-mzsx6"] Nov 25 10:18:56 crc kubenswrapper[4926]: I1125 10:18:56.025322 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec51083f-ad2b-4b95-ba9f-de0a8db553fc" path="/var/lib/kubelet/pods/ec51083f-ad2b-4b95-ba9f-de0a8db553fc/volumes" Nov 25 10:19:28 crc kubenswrapper[4926]: I1125 10:19:28.568061 4926 scope.go:117] "RemoveContainer" containerID="9b382af683c76723ffeebe2758d190df26f88b17ebe44ced6b91fc0841029d6f" Nov 25 10:21:14 crc kubenswrapper[4926]: I1125 10:21:14.018125 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:21:14 crc kubenswrapper[4926]: I1125 10:21:14.018720 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.862749 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 10:21:26 crc kubenswrapper[4926]: E1125 10:21:26.863877 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="registry-server" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.863901 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="registry-server" Nov 25 10:21:26 crc kubenswrapper[4926]: E1125 10:21:26.863968 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="registry-server" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.863980 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="registry-server" Nov 25 10:21:26 crc kubenswrapper[4926]: E1125 10:21:26.864078 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="extract-content" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.864095 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="extract-content" Nov 25 10:21:26 crc kubenswrapper[4926]: E1125 10:21:26.864117 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="extract-utilities" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.864130 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="extract-utilities" Nov 25 10:21:26 crc kubenswrapper[4926]: E1125 10:21:26.864153 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="extract-content" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.864164 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="extract-content" Nov 25 10:21:26 crc kubenswrapper[4926]: E1125 10:21:26.864194 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="extract-utilities" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.864206 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="extract-utilities" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.864509 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="05218437-b051-4be9-9a59-89a44ef50530" containerName="registry-server" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.864533 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fb1b1de-dfd5-4afb-b31b-45e396de9dc9" containerName="registry-server" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.865559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.873868 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.874133 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.874482 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.971011 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a986e749-6f64-454b-b2ef-a1438fee759d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:26 crc kubenswrapper[4926]: I1125 10:21:26.971134 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a986e749-6f64-454b-b2ef-a1438fee759d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:27 crc kubenswrapper[4926]: I1125 10:21:27.072649 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a986e749-6f64-454b-b2ef-a1438fee759d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:27 crc kubenswrapper[4926]: I1125 10:21:27.072728 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a986e749-6f64-454b-b2ef-a1438fee759d-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:27 crc kubenswrapper[4926]: I1125 10:21:27.073092 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a986e749-6f64-454b-b2ef-a1438fee759d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:27 crc kubenswrapper[4926]: I1125 10:21:27.095322 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a986e749-6f64-454b-b2ef-a1438fee759d-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:27 crc kubenswrapper[4926]: I1125 10:21:27.184729 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:27 crc kubenswrapper[4926]: I1125 10:21:27.629066 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 25 10:21:28 crc kubenswrapper[4926]: I1125 10:21:28.300798 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a986e749-6f64-454b-b2ef-a1438fee759d","Type":"ContainerStarted","Data":"b9f3f0f0681a8ae9e2fa9f45ebbfbcf812c50a35b109e9969e3f3fe183455c96"} Nov 25 10:21:28 crc kubenswrapper[4926]: I1125 10:21:28.301120 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a986e749-6f64-454b-b2ef-a1438fee759d","Type":"ContainerStarted","Data":"83b21e6c4942431a7a73a3294375e5a113d8015fa1451b9cbc7650f9d002e49e"} Nov 25 10:21:28 crc kubenswrapper[4926]: I1125 10:21:28.324989 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.324958969 podStartE2EDuration="2.324958969s" podCreationTimestamp="2025-11-25 10:21:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:21:28.317795441 +0000 UTC m=+1985.001751878" watchObservedRunningTime="2025-11-25 10:21:28.324958969 +0000 UTC m=+1985.008915436" Nov 25 10:21:29 crc kubenswrapper[4926]: I1125 10:21:29.327754 4926 generic.go:334] "Generic (PLEG): container finished" podID="a986e749-6f64-454b-b2ef-a1438fee759d" containerID="b9f3f0f0681a8ae9e2fa9f45ebbfbcf812c50a35b109e9969e3f3fe183455c96" exitCode=0 Nov 25 10:21:29 crc kubenswrapper[4926]: I1125 10:21:29.327805 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a986e749-6f64-454b-b2ef-a1438fee759d","Type":"ContainerDied","Data":"b9f3f0f0681a8ae9e2fa9f45ebbfbcf812c50a35b109e9969e3f3fe183455c96"} Nov 25 10:21:30 crc kubenswrapper[4926]: I1125 10:21:30.749995 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:30 crc kubenswrapper[4926]: I1125 10:21:30.945546 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a986e749-6f64-454b-b2ef-a1438fee759d-kube-api-access\") pod \"a986e749-6f64-454b-b2ef-a1438fee759d\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " Nov 25 10:21:30 crc kubenswrapper[4926]: I1125 10:21:30.946355 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a986e749-6f64-454b-b2ef-a1438fee759d-kubelet-dir\") pod \"a986e749-6f64-454b-b2ef-a1438fee759d\" (UID: \"a986e749-6f64-454b-b2ef-a1438fee759d\") " Nov 25 10:21:30 crc kubenswrapper[4926]: I1125 10:21:30.946463 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a986e749-6f64-454b-b2ef-a1438fee759d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a986e749-6f64-454b-b2ef-a1438fee759d" (UID: "a986e749-6f64-454b-b2ef-a1438fee759d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:21:30 crc kubenswrapper[4926]: I1125 10:21:30.947387 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a986e749-6f64-454b-b2ef-a1438fee759d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:21:30 crc kubenswrapper[4926]: I1125 10:21:30.950763 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a986e749-6f64-454b-b2ef-a1438fee759d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a986e749-6f64-454b-b2ef-a1438fee759d" (UID: "a986e749-6f64-454b-b2ef-a1438fee759d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:21:31 crc kubenswrapper[4926]: I1125 10:21:31.048829 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a986e749-6f64-454b-b2ef-a1438fee759d-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:21:31 crc kubenswrapper[4926]: I1125 10:21:31.349567 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"a986e749-6f64-454b-b2ef-a1438fee759d","Type":"ContainerDied","Data":"83b21e6c4942431a7a73a3294375e5a113d8015fa1451b9cbc7650f9d002e49e"} Nov 25 10:21:31 crc kubenswrapper[4926]: I1125 10:21:31.349607 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 25 10:21:31 crc kubenswrapper[4926]: I1125 10:21:31.349612 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="83b21e6c4942431a7a73a3294375e5a113d8015fa1451b9cbc7650f9d002e49e" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.061169 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 10:21:34 crc kubenswrapper[4926]: E1125 10:21:34.061552 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a986e749-6f64-454b-b2ef-a1438fee759d" containerName="pruner" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.061565 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a986e749-6f64-454b-b2ef-a1438fee759d" containerName="pruner" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.061778 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a986e749-6f64-454b-b2ef-a1438fee759d" containerName="pruner" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.062387 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.065621 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.065780 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.071192 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.201098 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.201476 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kube-api-access\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.201807 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-var-lock\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.304056 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-var-lock\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.304170 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.304174 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-var-lock\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.304327 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kubelet-dir\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.304380 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kube-api-access\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.329864 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kube-api-access\") pod \"installer-9-crc\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.382717 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:21:34 crc kubenswrapper[4926]: I1125 10:21:34.800453 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 25 10:21:35 crc kubenswrapper[4926]: I1125 10:21:35.389526 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bb9f2bfe-dd62-4c39-8864-f44219cfb999","Type":"ContainerStarted","Data":"76ca155cb5fae156bcfbc3bd8623e60e1659d03c3bb84495380e00789b133bcc"} Nov 25 10:21:35 crc kubenswrapper[4926]: I1125 10:21:35.389843 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bb9f2bfe-dd62-4c39-8864-f44219cfb999","Type":"ContainerStarted","Data":"5d7289affc096d9e1cd6d33a80416337962ea83931b4ebbee5e6368e79db94c6"} Nov 25 10:21:35 crc kubenswrapper[4926]: I1125 10:21:35.411294 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.411274564 podStartE2EDuration="1.411274564s" podCreationTimestamp="2025-11-25 10:21:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:21:35.404005093 +0000 UTC m=+1992.087961530" watchObservedRunningTime="2025-11-25 10:21:35.411274564 +0000 UTC m=+1992.095230991" Nov 25 10:21:44 crc kubenswrapper[4926]: I1125 10:21:44.021615 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:21:44 crc kubenswrapper[4926]: I1125 10:21:44.029419 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.315984 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hwnrz"] Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.318539 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.329390 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwnrz"] Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.451689 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-utilities\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.451965 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2trw8\" (UniqueName: \"kubernetes.io/projected/7b79e795-059c-4e8e-b19d-02698fc01cd4-kube-api-access-2trw8\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.452065 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-catalog-content\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.554668 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-catalog-content\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.554794 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-utilities\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.554919 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2trw8\" (UniqueName: \"kubernetes.io/projected/7b79e795-059c-4e8e-b19d-02698fc01cd4-kube-api-access-2trw8\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.555565 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-catalog-content\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.555602 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-utilities\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.576998 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2trw8\" (UniqueName: \"kubernetes.io/projected/7b79e795-059c-4e8e-b19d-02698fc01cd4-kube-api-access-2trw8\") pod \"redhat-marketplace-hwnrz\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:47 crc kubenswrapper[4926]: I1125 10:21:47.644830 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:48 crc kubenswrapper[4926]: I1125 10:21:48.171352 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwnrz"] Nov 25 10:21:48 crc kubenswrapper[4926]: I1125 10:21:48.489108 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerID="791f9e0a7ed5ac2e014c9d332708b46344de71c2f201455cc82902369d902211" exitCode=0 Nov 25 10:21:48 crc kubenswrapper[4926]: I1125 10:21:48.489230 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwnrz" event={"ID":"7b79e795-059c-4e8e-b19d-02698fc01cd4","Type":"ContainerDied","Data":"791f9e0a7ed5ac2e014c9d332708b46344de71c2f201455cc82902369d902211"} Nov 25 10:21:48 crc kubenswrapper[4926]: I1125 10:21:48.489260 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwnrz" event={"ID":"7b79e795-059c-4e8e-b19d-02698fc01cd4","Type":"ContainerStarted","Data":"a714dfaa25c28ce22869f81ae1aead642f12e29008673e9af1d0e5988697a83a"} Nov 25 10:21:48 crc kubenswrapper[4926]: I1125 10:21:48.501660 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:21:50 crc kubenswrapper[4926]: I1125 10:21:50.516718 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerID="130e616df46d11d674d551475914365c96f751e0cb5019a2e4a1326483d4b7f9" exitCode=0 Nov 25 10:21:50 crc kubenswrapper[4926]: I1125 10:21:50.517285 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwnrz" event={"ID":"7b79e795-059c-4e8e-b19d-02698fc01cd4","Type":"ContainerDied","Data":"130e616df46d11d674d551475914365c96f751e0cb5019a2e4a1326483d4b7f9"} Nov 25 10:21:51 crc kubenswrapper[4926]: I1125 10:21:51.529683 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwnrz" event={"ID":"7b79e795-059c-4e8e-b19d-02698fc01cd4","Type":"ContainerStarted","Data":"a162f4f5232e3c610b8f4571e5e097ee6c418ff2db81de5c34b03b5d7cf58aa2"} Nov 25 10:21:51 crc kubenswrapper[4926]: I1125 10:21:51.556731 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hwnrz" podStartSLOduration=1.7327624830000001 podStartE2EDuration="4.556713485s" podCreationTimestamp="2025-11-25 10:21:47 +0000 UTC" firstStartedPulling="2025-11-25 10:21:48.492227946 +0000 UTC m=+2005.176184373" lastFinishedPulling="2025-11-25 10:21:51.316178918 +0000 UTC m=+2008.000135375" observedRunningTime="2025-11-25 10:21:51.549646679 +0000 UTC m=+2008.233603116" watchObservedRunningTime="2025-11-25 10:21:51.556713485 +0000 UTC m=+2008.240669912" Nov 25 10:21:57 crc kubenswrapper[4926]: I1125 10:21:57.645879 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:57 crc kubenswrapper[4926]: I1125 10:21:57.646372 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:57 crc kubenswrapper[4926]: I1125 10:21:57.689098 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:58 crc kubenswrapper[4926]: I1125 10:21:58.629911 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:21:58 crc kubenswrapper[4926]: I1125 10:21:58.673640 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwnrz"] Nov 25 10:22:00 crc kubenswrapper[4926]: I1125 10:22:00.652505 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hwnrz" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="registry-server" containerID="cri-o://a162f4f5232e3c610b8f4571e5e097ee6c418ff2db81de5c34b03b5d7cf58aa2" gracePeriod=2 Nov 25 10:22:01 crc kubenswrapper[4926]: I1125 10:22:01.661089 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerID="a162f4f5232e3c610b8f4571e5e097ee6c418ff2db81de5c34b03b5d7cf58aa2" exitCode=0 Nov 25 10:22:01 crc kubenswrapper[4926]: I1125 10:22:01.661406 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwnrz" event={"ID":"7b79e795-059c-4e8e-b19d-02698fc01cd4","Type":"ContainerDied","Data":"a162f4f5232e3c610b8f4571e5e097ee6c418ff2db81de5c34b03b5d7cf58aa2"} Nov 25 10:22:01 crc kubenswrapper[4926]: I1125 10:22:01.946205 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.106696 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-catalog-content\") pod \"7b79e795-059c-4e8e-b19d-02698fc01cd4\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.106810 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-utilities\") pod \"7b79e795-059c-4e8e-b19d-02698fc01cd4\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.107002 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2trw8\" (UniqueName: \"kubernetes.io/projected/7b79e795-059c-4e8e-b19d-02698fc01cd4-kube-api-access-2trw8\") pod \"7b79e795-059c-4e8e-b19d-02698fc01cd4\" (UID: \"7b79e795-059c-4e8e-b19d-02698fc01cd4\") " Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.108704 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-utilities" (OuterVolumeSpecName: "utilities") pod "7b79e795-059c-4e8e-b19d-02698fc01cd4" (UID: "7b79e795-059c-4e8e-b19d-02698fc01cd4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.126078 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7b79e795-059c-4e8e-b19d-02698fc01cd4" (UID: "7b79e795-059c-4e8e-b19d-02698fc01cd4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.128091 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b79e795-059c-4e8e-b19d-02698fc01cd4-kube-api-access-2trw8" (OuterVolumeSpecName: "kube-api-access-2trw8") pod "7b79e795-059c-4e8e-b19d-02698fc01cd4" (UID: "7b79e795-059c-4e8e-b19d-02698fc01cd4"). InnerVolumeSpecName "kube-api-access-2trw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.215405 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2trw8\" (UniqueName: \"kubernetes.io/projected/7b79e795-059c-4e8e-b19d-02698fc01cd4-kube-api-access-2trw8\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.215441 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.215450 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7b79e795-059c-4e8e-b19d-02698fc01cd4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.674525 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hwnrz" event={"ID":"7b79e795-059c-4e8e-b19d-02698fc01cd4","Type":"ContainerDied","Data":"a714dfaa25c28ce22869f81ae1aead642f12e29008673e9af1d0e5988697a83a"} Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.674587 4926 scope.go:117] "RemoveContainer" containerID="a162f4f5232e3c610b8f4571e5e097ee6c418ff2db81de5c34b03b5d7cf58aa2" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.674609 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hwnrz" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.705581 4926 scope.go:117] "RemoveContainer" containerID="130e616df46d11d674d551475914365c96f751e0cb5019a2e4a1326483d4b7f9" Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.715347 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwnrz"] Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.727857 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hwnrz"] Nov 25 10:22:02 crc kubenswrapper[4926]: I1125 10:22:02.735284 4926 scope.go:117] "RemoveContainer" containerID="791f9e0a7ed5ac2e014c9d332708b46344de71c2f201455cc82902369d902211" Nov 25 10:22:04 crc kubenswrapper[4926]: I1125 10:22:04.021340 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" path="/var/lib/kubelet/pods/7b79e795-059c-4e8e-b19d-02698fc01cd4/volumes" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.874784 4926 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.875698 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="extract-utilities" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.875712 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="extract-utilities" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.875733 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="registry-server" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.875740 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="registry-server" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.875767 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="extract-content" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.875774 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="extract-content" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.875991 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b79e795-059c-4e8e-b19d-02698fc01cd4" containerName="registry-server" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.876697 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.916522 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.953650 4926 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.954200 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9" gracePeriod=15 Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.954258 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7" gracePeriod=15 Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.954298 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99" gracePeriod=15 Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.954347 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48" gracePeriod=15 Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.954244 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530" gracePeriod=15 Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955084 4926 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.955574 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955602 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.955614 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955623 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.955656 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955664 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.955685 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955693 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.955716 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955723 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.955741 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955749 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.955983 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.956002 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.956013 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.956021 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.956036 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 25 10:22:12 crc kubenswrapper[4926]: E1125 10:22:12.956242 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.956258 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:22:12 crc kubenswrapper[4926]: I1125 10:22:12.956474 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.012466 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.012584 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.012645 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.012697 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.012725 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.114832 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.114886 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.114909 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.114967 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115115 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115444 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115495 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115544 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115594 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115652 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115880 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115971 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.115998 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: E1125 10:22:13.117243 4926 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.90:6443: connect: connection refused" pod="openshift-image-registry/image-registry-66df7c8f76-zvms9" volumeName="registry-storage" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.208702 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.219813 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.219849 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.219895 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.220481 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.220553 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.220613 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:13 crc kubenswrapper[4926]: E1125 10:22:13.271928 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.90:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b38cd41eec162 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:22:13.271298402 +0000 UTC m=+2029.955254839,LastTimestamp:2025-11-25 10:22:13.271298402 +0000 UTC m=+2029.955254839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.769727 4926 generic.go:334] "Generic (PLEG): container finished" podID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" containerID="76ca155cb5fae156bcfbc3bd8623e60e1659d03c3bb84495380e00789b133bcc" exitCode=0 Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.769815 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bb9f2bfe-dd62-4c39-8864-f44219cfb999","Type":"ContainerDied","Data":"76ca155cb5fae156bcfbc3bd8623e60e1659d03c3bb84495380e00789b133bcc"} Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.771390 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.771881 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.772358 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.772497 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f"} Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.772535 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c7578ebea3c9bfa54abb9df0ce0a2f3e0c8d8378c45b035da2ee0d3aebc82002"} Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.773388 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.773677 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.773977 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.775180 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.776656 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.777410 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530" exitCode=0 Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.777522 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48" exitCode=0 Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.777609 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7" exitCode=0 Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.777691 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99" exitCode=2 Nov 25 10:22:13 crc kubenswrapper[4926]: I1125 10:22:13.777464 4926 scope.go:117] "RemoveContainer" containerID="52eef3e90440ef6508ef145d12c09d422f5d7d5f2e7b71d24e512112030c95be" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.014584 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.014628 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.020972 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.021298 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.021635 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.022261 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.022821 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6f381f071189398b29183f9732862d78affd0630860a5419ed3e39c642baae0"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.022872 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://d6f381f071189398b29183f9732862d78affd0630860a5419ed3e39c642baae0" gracePeriod=600 Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.796512 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.801641 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="d6f381f071189398b29183f9732862d78affd0630860a5419ed3e39c642baae0" exitCode=0 Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.801779 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"d6f381f071189398b29183f9732862d78affd0630860a5419ed3e39c642baae0"} Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.801869 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28"} Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.801917 4926 scope.go:117] "RemoveContainer" containerID="5602b18b3635474390795a7fdfe2fe9ec24ba9025dad4bcb2955505d43689625" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.802835 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.803338 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:14 crc kubenswrapper[4926]: I1125 10:22:14.804279 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.240720 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.242045 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.242549 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.242827 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.351257 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kube-api-access\") pod \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.351392 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kubelet-dir\") pod \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.351454 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-var-lock\") pod \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\" (UID: \"bb9f2bfe-dd62-4c39-8864-f44219cfb999\") " Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.351676 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bb9f2bfe-dd62-4c39-8864-f44219cfb999" (UID: "bb9f2bfe-dd62-4c39-8864-f44219cfb999"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.352098 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-var-lock" (OuterVolumeSpecName: "var-lock") pod "bb9f2bfe-dd62-4c39-8864-f44219cfb999" (UID: "bb9f2bfe-dd62-4c39-8864-f44219cfb999"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.352310 4926 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.352328 4926 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/bb9f2bfe-dd62-4c39-8864-f44219cfb999-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.357918 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bb9f2bfe-dd62-4c39-8864-f44219cfb999" (UID: "bb9f2bfe-dd62-4c39-8864-f44219cfb999"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.453641 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bb9f2bfe-dd62-4c39-8864-f44219cfb999-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.467833 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.468960 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.469652 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.470340 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.470945 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.471252 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.498975 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.655930 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.656011 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.656391 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.656519 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.656693 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.656881 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.657194 4926 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.657258 4926 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.657312 4926 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.812094 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.812868 4926 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9" exitCode=0 Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.813003 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.813016 4926 scope.go:117] "RemoveContainer" containerID="55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.814706 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.815003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"bb9f2bfe-dd62-4c39-8864-f44219cfb999","Type":"ContainerDied","Data":"5d7289affc096d9e1cd6d33a80416337962ea83931b4ebbee5e6368e79db94c6"} Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.815047 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d7289affc096d9e1cd6d33a80416337962ea83931b4ebbee5e6368e79db94c6" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.837145 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.837529 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.838031 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.838458 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.838882 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.839286 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.839492 4926 scope.go:117] "RemoveContainer" containerID="21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.839604 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.840182 4926 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.864466 4926 scope.go:117] "RemoveContainer" containerID="73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.882811 4926 scope.go:117] "RemoveContainer" containerID="4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.915635 4926 scope.go:117] "RemoveContainer" containerID="203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.942656 4926 scope.go:117] "RemoveContainer" containerID="5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.975650 4926 scope.go:117] "RemoveContainer" containerID="55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530" Nov 25 10:22:15 crc kubenswrapper[4926]: E1125 10:22:15.976664 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\": container with ID starting with 55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530 not found: ID does not exist" containerID="55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.976713 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530"} err="failed to get container status \"55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\": rpc error: code = NotFound desc = could not find container \"55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530\": container with ID starting with 55fe5630e62858188d78e81b5201db38d09eb9e92a1eeaa9f77cf2b93bdb8530 not found: ID does not exist" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.976738 4926 scope.go:117] "RemoveContainer" containerID="21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48" Nov 25 10:22:15 crc kubenswrapper[4926]: E1125 10:22:15.979814 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\": container with ID starting with 21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48 not found: ID does not exist" containerID="21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.979871 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48"} err="failed to get container status \"21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\": rpc error: code = NotFound desc = could not find container \"21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48\": container with ID starting with 21c445c470577169f79d2539e7047deb0e174d0bb22eef579b3c22f7ede2fb48 not found: ID does not exist" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.979890 4926 scope.go:117] "RemoveContainer" containerID="73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7" Nov 25 10:22:15 crc kubenswrapper[4926]: E1125 10:22:15.980414 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\": container with ID starting with 73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7 not found: ID does not exist" containerID="73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.980465 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7"} err="failed to get container status \"73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\": rpc error: code = NotFound desc = could not find container \"73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7\": container with ID starting with 73c49e56f59a50ae4506ad9d937e732878448b19ee51faad033092a2901b4cb7 not found: ID does not exist" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.980503 4926 scope.go:117] "RemoveContainer" containerID="4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99" Nov 25 10:22:15 crc kubenswrapper[4926]: E1125 10:22:15.982052 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\": container with ID starting with 4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99 not found: ID does not exist" containerID="4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.982099 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99"} err="failed to get container status \"4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\": rpc error: code = NotFound desc = could not find container \"4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99\": container with ID starting with 4524e32a0d9b29d3407297e8396d8dd29288a809cac33d1294dc7fd05d7c7e99 not found: ID does not exist" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.982123 4926 scope.go:117] "RemoveContainer" containerID="203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9" Nov 25 10:22:15 crc kubenswrapper[4926]: E1125 10:22:15.982512 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\": container with ID starting with 203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9 not found: ID does not exist" containerID="203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.982539 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9"} err="failed to get container status \"203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\": rpc error: code = NotFound desc = could not find container \"203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9\": container with ID starting with 203d2c0341d922b48ad700b8fe526ea7f034f76e851a1f473d71b16ebf3331f9 not found: ID does not exist" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.982558 4926 scope.go:117] "RemoveContainer" containerID="5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8" Nov 25 10:22:15 crc kubenswrapper[4926]: E1125 10:22:15.982822 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\": container with ID starting with 5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8 not found: ID does not exist" containerID="5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8" Nov 25 10:22:15 crc kubenswrapper[4926]: I1125 10:22:15.982847 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8"} err="failed to get container status \"5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\": rpc error: code = NotFound desc = could not find container \"5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8\": container with ID starting with 5898e4e348d34e35a79dc0139b00ed2b76c13bfd9ade8fcb9e0b2ec379564ec8 not found: ID does not exist" Nov 25 10:22:16 crc kubenswrapper[4926]: I1125 10:22:16.043162 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 25 10:22:17 crc kubenswrapper[4926]: E1125 10:22:17.035502 4926 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.90:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187b38cd41eec162 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-25 10:22:13.271298402 +0000 UTC m=+2029.955254839,LastTimestamp:2025-11-25 10:22:13.271298402 +0000 UTC m=+2029.955254839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.675119 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.676242 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.676694 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.676923 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.677270 4926 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:19 crc kubenswrapper[4926]: I1125 10:22:19.677300 4926 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.677574 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="200ms" Nov 25 10:22:19 crc kubenswrapper[4926]: E1125 10:22:19.878716 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="400ms" Nov 25 10:22:20 crc kubenswrapper[4926]: E1125 10:22:20.279997 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="800ms" Nov 25 10:22:21 crc kubenswrapper[4926]: E1125 10:22:21.081109 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="1.6s" Nov 25 10:22:22 crc kubenswrapper[4926]: E1125 10:22:22.681773 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="3.2s" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.022707 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.024413 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.024973 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.905543 4926 generic.go:334] "Generic (PLEG): container finished" podID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" containerID="0580482c8e707e87bce39e86dd5a0fc30e705dc66256d88ecce3aab5b1b165e4" exitCode=1 Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.905599 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" event={"ID":"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef","Type":"ContainerDied","Data":"0580482c8e707e87bce39e86dd5a0fc30e705dc66256d88ecce3aab5b1b165e4"} Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.906497 4926 scope.go:117] "RemoveContainer" containerID="0580482c8e707e87bce39e86dd5a0fc30e705dc66256d88ecce3aab5b1b165e4" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.906677 4926 status_manager.go:851] "Failed to get status for pod" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-65f5fc5bb5-r6kqg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.906855 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.907096 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:24 crc kubenswrapper[4926]: I1125 10:22:24.907618 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.011324 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.012404 4926 status_manager.go:851] "Failed to get status for pod" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-65f5fc5bb5-r6kqg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.012658 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.013041 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.013520 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.033716 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.033748 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:25 crc kubenswrapper[4926]: E1125 10:22:25.034279 4926 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.035045 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:25 crc kubenswrapper[4926]: W1125 10:22:25.076807 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-48ae8b372019d44ed747eb26caed17f0303556be3276888c7013068b47fa5260 WatchSource:0}: Error finding container 48ae8b372019d44ed747eb26caed17f0303556be3276888c7013068b47fa5260: Status 404 returned error can't find the container with id 48ae8b372019d44ed747eb26caed17f0303556be3276888c7013068b47fa5260 Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.499212 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 10:22:25 crc kubenswrapper[4926]: E1125 10:22:25.883969 4926 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.90:6443: connect: connection refused" interval="6.4s" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.923488 4926 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="8905584e89cfb7e902ffa7866bbc4cfa2796d9e09ce888f444bbed34f2f30d81" exitCode=0 Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.923579 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"8905584e89cfb7e902ffa7866bbc4cfa2796d9e09ce888f444bbed34f2f30d81"} Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.923630 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"48ae8b372019d44ed747eb26caed17f0303556be3276888c7013068b47fa5260"} Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.923961 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.923980 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:25 crc kubenswrapper[4926]: E1125 10:22:25.924270 4926 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.924795 4926 status_manager.go:851] "Failed to get status for pod" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-65f5fc5bb5-r6kqg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.925205 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.925721 4926 generic.go:334] "Generic (PLEG): container finished" podID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" containerID="4d63c5b88246faa5057a3d692afb7f7d47cae3af0cdc6903a4a2a80ec084bbb7" exitCode=1 Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.925742 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" event={"ID":"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef","Type":"ContainerDied","Data":"4d63c5b88246faa5057a3d692afb7f7d47cae3af0cdc6903a4a2a80ec084bbb7"} Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.925762 4926 scope.go:117] "RemoveContainer" containerID="0580482c8e707e87bce39e86dd5a0fc30e705dc66256d88ecce3aab5b1b165e4" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.926132 4926 scope.go:117] "RemoveContainer" containerID="4d63c5b88246faa5057a3d692afb7f7d47cae3af0cdc6903a4a2a80ec084bbb7" Nov 25 10:22:25 crc kubenswrapper[4926]: E1125 10:22:25.926307 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-65f5fc5bb5-r6kqg_metallb-system(0cb1f5f8-6fcb-4664-8fcc-74e416d63eef)\"" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.927173 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.927379 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.927811 4926 status_manager.go:851] "Failed to get status for pod" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.928674 4926 status_manager.go:851] "Failed to get status for pod" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-65f5fc5bb5-r6kqg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.929369 4926 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:25 crc kubenswrapper[4926]: I1125 10:22:25.929801 4926 status_manager.go:851] "Failed to get status for pod" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/pods/machine-config-daemon-b82dg\": dial tcp 38.129.56.90:6443: connect: connection refused" Nov 25 10:22:26 crc kubenswrapper[4926]: I1125 10:22:26.938243 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"47d9a3a0ec7cbf3e4a4aeb6ddff68f7ff377b4cfdbff1268cae5a6780d02856e"} Nov 25 10:22:26 crc kubenswrapper[4926]: I1125 10:22:26.938765 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"66b45d072385a6e4fcc5a103206a1ba09afccb8774298f8c5818970465f7524f"} Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.457077 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.457919 4926 scope.go:117] "RemoveContainer" containerID="4d63c5b88246faa5057a3d692afb7f7d47cae3af0cdc6903a4a2a80ec084bbb7" Nov 25 10:22:27 crc kubenswrapper[4926]: E1125 10:22:27.458229 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-65f5fc5bb5-r6kqg_metallb-system(0cb1f5f8-6fcb-4664-8fcc-74e416d63eef)\"" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.952358 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.952664 4926 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae" exitCode=1 Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.952730 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae"} Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.953472 4926 scope.go:117] "RemoveContainer" containerID="68d09294949d8e8ad9138577142fff229f6e3e19fc1e3947a143a65eacf937ae" Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.956275 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bfbfbad82b49e0bb19b30cfbaf9a7defadad3a50bb1e5302b8dc467b3f491c35"} Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.956317 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"35395ad724d984929cb0d525f1d3841c3a12e158f96ec426f37b2d5dca9c13b3"} Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.956328 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"bfe00cda10a5456b72e76ec55eb7fcf898ae8f0001d42ffb900679383fbc00ce"} Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.956453 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.956576 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:27 crc kubenswrapper[4926]: I1125 10:22:27.956591 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:28 crc kubenswrapper[4926]: I1125 10:22:28.862271 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:22:28 crc kubenswrapper[4926]: I1125 10:22:28.972849 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 25 10:22:28 crc kubenswrapper[4926]: I1125 10:22:28.972902 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"418c6e021ba45037fbe3bc0740bd0a98e4b4094b85ede6cfa86e2762a4e856aa"} Nov 25 10:22:30 crc kubenswrapper[4926]: I1125 10:22:30.035269 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:30 crc kubenswrapper[4926]: I1125 10:22:30.035310 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:30 crc kubenswrapper[4926]: I1125 10:22:30.040833 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:32 crc kubenswrapper[4926]: I1125 10:22:32.964625 4926 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:33 crc kubenswrapper[4926]: I1125 10:22:33.002179 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:33 crc kubenswrapper[4926]: I1125 10:22:33.002208 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:33 crc kubenswrapper[4926]: I1125 10:22:33.006306 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.028302 4926 generic.go:334] "Generic (PLEG): container finished" podID="5cf84563-5163-4108-92dc-1dbf757212de" containerID="720748cdb5ddb350e01205c49d4237fe193664c59c467ad1a9e97aa1241d74df" exitCode=1 Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.034243 4926 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.034274 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="b76756dc-1402-4054-bff2-e76ba1281b6f" Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.034341 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" event={"ID":"5cf84563-5163-4108-92dc-1dbf757212de","Type":"ContainerDied","Data":"720748cdb5ddb350e01205c49d4237fe193664c59c467ad1a9e97aa1241d74df"} Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.034884 4926 scope.go:117] "RemoveContainer" containerID="720748cdb5ddb350e01205c49d4237fe193664c59c467ad1a9e97aa1241d74df" Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.048604 4926 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="0b910686-3061-413c-bc27-2064f97aca55" Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.121317 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:22:34 crc kubenswrapper[4926]: I1125 10:22:34.326655 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" podUID="2494b104-d923-4de0-82e0-8f4ff0e9c5f8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.93:8081/readyz\": dial tcp 10.217.0.93:8081: connect: connection refused" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.039503 4926 generic.go:334] "Generic (PLEG): container finished" podID="db149e24-86d6-4f34-abb0-82229e25e9d7" containerID="5efd45a115f512a93e860eca1c1f1d00774a08e91e9be3c03ebf65fecffd6f90" exitCode=1 Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.040425 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" event={"ID":"db149e24-86d6-4f34-abb0-82229e25e9d7","Type":"ContainerDied","Data":"5efd45a115f512a93e860eca1c1f1d00774a08e91e9be3c03ebf65fecffd6f90"} Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.041095 4926 scope.go:117] "RemoveContainer" containerID="5efd45a115f512a93e860eca1c1f1d00774a08e91e9be3c03ebf65fecffd6f90" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.041165 4926 generic.go:334] "Generic (PLEG): container finished" podID="b6772bac-4625-41d7-915d-2fdabd9916b7" containerID="10425a230b9a32fe43c39a81bf425675d2869d9c23771ebaafa529e3812ee55e" exitCode=1 Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.041220 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" event={"ID":"b6772bac-4625-41d7-915d-2fdabd9916b7","Type":"ContainerDied","Data":"10425a230b9a32fe43c39a81bf425675d2869d9c23771ebaafa529e3812ee55e"} Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.041754 4926 scope.go:117] "RemoveContainer" containerID="10425a230b9a32fe43c39a81bf425675d2869d9c23771ebaafa529e3812ee55e" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.044218 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" event={"ID":"5cf84563-5163-4108-92dc-1dbf757212de","Type":"ContainerStarted","Data":"f2679d4e4060aa75eca78be93552f3bcc2414c196906b5e302d6c7cd802e569e"} Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.044467 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.058465 4926 generic.go:334] "Generic (PLEG): container finished" podID="edb138df-3c28-4b7d-b285-2fe43094ead3" containerID="1fff1d775d23d3f0cc426529f74c311a05723ee591a8f9e68adf4b7f0899a62f" exitCode=1 Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.058537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" event={"ID":"edb138df-3c28-4b7d-b285-2fe43094ead3","Type":"ContainerDied","Data":"1fff1d775d23d3f0cc426529f74c311a05723ee591a8f9e68adf4b7f0899a62f"} Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.059241 4926 scope.go:117] "RemoveContainer" containerID="1fff1d775d23d3f0cc426529f74c311a05723ee591a8f9e68adf4b7f0899a62f" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.062690 4926 generic.go:334] "Generic (PLEG): container finished" podID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" containerID="ed19bb0315e4fa43211796d7d47cb107e576e80156b84d62a677779e9547cb96" exitCode=1 Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.062748 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" event={"ID":"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8","Type":"ContainerDied","Data":"ed19bb0315e4fa43211796d7d47cb107e576e80156b84d62a677779e9547cb96"} Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.063384 4926 scope.go:117] "RemoveContainer" containerID="ed19bb0315e4fa43211796d7d47cb107e576e80156b84d62a677779e9547cb96" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.065371 4926 generic.go:334] "Generic (PLEG): container finished" podID="2494b104-d923-4de0-82e0-8f4ff0e9c5f8" containerID="77d1e8202eba21168861272b6c126891e870b62c52177ff38897d61849fed639" exitCode=1 Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.065505 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" event={"ID":"2494b104-d923-4de0-82e0-8f4ff0e9c5f8","Type":"ContainerDied","Data":"77d1e8202eba21168861272b6c126891e870b62c52177ff38897d61849fed639"} Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.067624 4926 scope.go:117] "RemoveContainer" containerID="77d1e8202eba21168861272b6c126891e870b62c52177ff38897d61849fed639" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.360816 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.499407 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerName="kube-state-metrics" probeResult="failure" output="HTTP probe failed with statuscode: 503" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.499703 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.500418 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"6fc501f19762f6105670f5e405746e249f6b955c4433daa84248acf85c9e864b"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Nov 25 10:22:35 crc kubenswrapper[4926]: I1125 10:22:35.500457 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerName="kube-state-metrics" containerID="cri-o://6fc501f19762f6105670f5e405746e249f6b955c4433daa84248acf85c9e864b" gracePeriod=30 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.075588 4926 generic.go:334] "Generic (PLEG): container finished" podID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" containerID="9fcc3ce783d8ba3ae2c657c88c82388db47bbbb8acf55e583dcf67a89db85724" exitCode=1 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.075679 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" event={"ID":"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e","Type":"ContainerDied","Data":"9fcc3ce783d8ba3ae2c657c88c82388db47bbbb8acf55e583dcf67a89db85724"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.077693 4926 scope.go:117] "RemoveContainer" containerID="9fcc3ce783d8ba3ae2c657c88c82388db47bbbb8acf55e583dcf67a89db85724" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.079127 4926 generic.go:334] "Generic (PLEG): container finished" podID="03cc50ce-1146-4953-901e-19d5bd7c468e" containerID="0017977077d412e6e68756b8087f27244ca4dd25c08e12d1c1a6d9f6b09ce1c1" exitCode=1 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.079193 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" event={"ID":"03cc50ce-1146-4953-901e-19d5bd7c468e","Type":"ContainerDied","Data":"0017977077d412e6e68756b8087f27244ca4dd25c08e12d1c1a6d9f6b09ce1c1"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.081429 4926 scope.go:117] "RemoveContainer" containerID="0017977077d412e6e68756b8087f27244ca4dd25c08e12d1c1a6d9f6b09ce1c1" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.084493 4926 generic.go:334] "Generic (PLEG): container finished" podID="edb138df-3c28-4b7d-b285-2fe43094ead3" containerID="c90c02ed94327d5dcc6994a7fa3e6e6d37115fbfc0e824c591c37fbae1eb196e" exitCode=1 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.084846 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" event={"ID":"edb138df-3c28-4b7d-b285-2fe43094ead3","Type":"ContainerDied","Data":"c90c02ed94327d5dcc6994a7fa3e6e6d37115fbfc0e824c591c37fbae1eb196e"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.084879 4926 scope.go:117] "RemoveContainer" containerID="1fff1d775d23d3f0cc426529f74c311a05723ee591a8f9e68adf4b7f0899a62f" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.085324 4926 scope.go:117] "RemoveContainer" containerID="c90c02ed94327d5dcc6994a7fa3e6e6d37115fbfc0e824c591c37fbae1eb196e" Nov 25 10:22:36 crc kubenswrapper[4926]: E1125 10:22:36.085536 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-hr2gx_openstack-operators(edb138df-3c28-4b7d-b285-2fe43094ead3)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" podUID="edb138df-3c28-4b7d-b285-2fe43094ead3" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.091773 4926 generic.go:334] "Generic (PLEG): container finished" podID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" containerID="c8a513033d37f8bbe12f05de8260056f80f8b11aca124b6c67eb363a21f060c5" exitCode=1 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.091862 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" event={"ID":"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8","Type":"ContainerDied","Data":"c8a513033d37f8bbe12f05de8260056f80f8b11aca124b6c67eb363a21f060c5"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.092825 4926 scope.go:117] "RemoveContainer" containerID="c8a513033d37f8bbe12f05de8260056f80f8b11aca124b6c67eb363a21f060c5" Nov 25 10:22:36 crc kubenswrapper[4926]: E1125 10:22:36.093186 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-tmxx7_openstack-operators(fa24e0a8-f9a4-4160-bac5-4d73e8579fd8)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" podUID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.095520 4926 generic.go:334] "Generic (PLEG): container finished" podID="194893a3-1c67-4a28-a67e-7d7eeef304a7" containerID="cffdc6c8d780ff8826404234f8d89e8422500931f06a3fc4b92d62a5484033b9" exitCode=1 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.095589 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" event={"ID":"194893a3-1c67-4a28-a67e-7d7eeef304a7","Type":"ContainerDied","Data":"cffdc6c8d780ff8826404234f8d89e8422500931f06a3fc4b92d62a5484033b9"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.096228 4926 scope.go:117] "RemoveContainer" containerID="cffdc6c8d780ff8826404234f8d89e8422500931f06a3fc4b92d62a5484033b9" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.098047 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" event={"ID":"2494b104-d923-4de0-82e0-8f4ff0e9c5f8","Type":"ContainerStarted","Data":"eb16d2a8ba390768e62134bc53f81f2944fc5d777ffb4561f23de1771d3b3806"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.098257 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.103890 4926 generic.go:334] "Generic (PLEG): container finished" podID="b6772bac-4625-41d7-915d-2fdabd9916b7" containerID="413682c48013fdeed4fc2d5b9656cf12a302415171973b50a7bde547dee98ca6" exitCode=1 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.104135 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" event={"ID":"b6772bac-4625-41d7-915d-2fdabd9916b7","Type":"ContainerDied","Data":"413682c48013fdeed4fc2d5b9656cf12a302415171973b50a7bde547dee98ca6"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.104825 4926 scope.go:117] "RemoveContainer" containerID="413682c48013fdeed4fc2d5b9656cf12a302415171973b50a7bde547dee98ca6" Nov 25 10:22:36 crc kubenswrapper[4926]: E1125 10:22:36.105272 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-858778c9dc-lwgz6_openstack-operators(b6772bac-4625-41d7-915d-2fdabd9916b7)\"" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podUID="b6772bac-4625-41d7-915d-2fdabd9916b7" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.109272 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerID="6fc501f19762f6105670f5e405746e249f6b955c4433daa84248acf85c9e864b" exitCode=2 Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.109333 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerDied","Data":"6fc501f19762f6105670f5e405746e249f6b955c4433daa84248acf85c9e864b"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.113702 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" event={"ID":"db149e24-86d6-4f34-abb0-82229e25e9d7","Type":"ContainerStarted","Data":"59235d797bc70a4fc439fb54853b328af43b133b304a40fb04647a52c73eb796"} Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.113972 4926 status_manager.go:317] "Container readiness changed for unknown container" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" containerID="cri-o://5efd45a115f512a93e860eca1c1f1d00774a08e91e9be3c03ebf65fecffd6f90" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.113996 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.426228 4926 scope.go:117] "RemoveContainer" containerID="ed19bb0315e4fa43211796d7d47cb107e576e80156b84d62a677779e9547cb96" Nov 25 10:22:36 crc kubenswrapper[4926]: I1125 10:22:36.808553 4926 scope.go:117] "RemoveContainer" containerID="10425a230b9a32fe43c39a81bf425675d2869d9c23771ebaafa529e3812ee55e" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.124378 4926 generic.go:334] "Generic (PLEG): container finished" podID="f74db279-9026-4869-a147-d45bf581b618" containerID="72594fd7a747457517e5f0c4214007b015ca36bca78226cee674fc1e7d3699f7" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.124478 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" event={"ID":"f74db279-9026-4869-a147-d45bf581b618","Type":"ContainerDied","Data":"72594fd7a747457517e5f0c4214007b015ca36bca78226cee674fc1e7d3699f7"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.126818 4926 scope.go:117] "RemoveContainer" containerID="72594fd7a747457517e5f0c4214007b015ca36bca78226cee674fc1e7d3699f7" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.127124 4926 generic.go:334] "Generic (PLEG): container finished" podID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" containerID="a323101e1d43b04f7a693896c68109ac61b3f975edf283514ae070693da13635" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.127212 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerDied","Data":"a323101e1d43b04f7a693896c68109ac61b3f975edf283514ae070693da13635"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.128037 4926 scope.go:117] "RemoveContainer" containerID="a323101e1d43b04f7a693896c68109ac61b3f975edf283514ae070693da13635" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.132280 4926 generic.go:334] "Generic (PLEG): container finished" podID="416161f0-e9fb-4909-83c0-dfe7ad15885f" containerID="fd3404684a77a8cab7d504c74b9a9f632755c7c1c63bff3fc75216a3e2b5c822" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.132350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" event={"ID":"416161f0-e9fb-4909-83c0-dfe7ad15885f","Type":"ContainerDied","Data":"fd3404684a77a8cab7d504c74b9a9f632755c7c1c63bff3fc75216a3e2b5c822"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.132729 4926 scope.go:117] "RemoveContainer" containerID="fd3404684a77a8cab7d504c74b9a9f632755c7c1c63bff3fc75216a3e2b5c822" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.167502 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae1df4bc-d341-49fd-a49a-e02b4d8714d1" containerID="8b6b76df23ca2d547afa40eacc6f3bce4d70e0b3f29f406496c7f0cb25b87a79" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.167689 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerDied","Data":"8b6b76df23ca2d547afa40eacc6f3bce4d70e0b3f29f406496c7f0cb25b87a79"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.168713 4926 scope.go:117] "RemoveContainer" containerID="8b6b76df23ca2d547afa40eacc6f3bce4d70e0b3f29f406496c7f0cb25b87a79" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.174500 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerStarted","Data":"f7956046f9d7cb816c6d7cf6ab39281c090e0f105f0b1f059972ec2e65667060"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.174588 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.178270 4926 generic.go:334] "Generic (PLEG): container finished" podID="5bf64213-939f-4f49-9030-9f9dcc78f8bb" containerID="3b6d22f86522965bb63d95715cfe147c87133d7900a3fc98424c22ab8736cb72" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.178420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerDied","Data":"3b6d22f86522965bb63d95715cfe147c87133d7900a3fc98424c22ab8736cb72"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.179198 4926 scope.go:117] "RemoveContainer" containerID="3b6d22f86522965bb63d95715cfe147c87133d7900a3fc98424c22ab8736cb72" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.180798 4926 generic.go:334] "Generic (PLEG): container finished" podID="03cc50ce-1146-4953-901e-19d5bd7c468e" containerID="2999aabfa41dbf288b61da882776f755069cdfa4ba9aac180b62ddc64f4b55ec" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.180853 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" event={"ID":"03cc50ce-1146-4953-901e-19d5bd7c468e","Type":"ContainerDied","Data":"2999aabfa41dbf288b61da882776f755069cdfa4ba9aac180b62ddc64f4b55ec"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.180881 4926 scope.go:117] "RemoveContainer" containerID="0017977077d412e6e68756b8087f27244ca4dd25c08e12d1c1a6d9f6b09ce1c1" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.181261 4926 scope.go:117] "RemoveContainer" containerID="2999aabfa41dbf288b61da882776f755069cdfa4ba9aac180b62ddc64f4b55ec" Nov 25 10:22:37 crc kubenswrapper[4926]: E1125 10:22:37.181466 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-zdqj6_openstack-operators(03cc50ce-1146-4953-901e-19d5bd7c468e)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podUID="03cc50ce-1146-4953-901e-19d5bd7c468e" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.185477 4926 generic.go:334] "Generic (PLEG): container finished" podID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" containerID="9dd1498849e62796c39436f05c633b26a66126e317de44db35d715b1d6a23a22" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.185563 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerDied","Data":"9dd1498849e62796c39436f05c633b26a66126e317de44db35d715b1d6a23a22"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.186324 4926 scope.go:117] "RemoveContainer" containerID="9dd1498849e62796c39436f05c633b26a66126e317de44db35d715b1d6a23a22" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.188228 4926 generic.go:334] "Generic (PLEG): container finished" podID="1154878f-e8a5-42de-916f-73276b979e74" containerID="a255b4fc4d71a17bfa31390573549daada5264701abd9f080f44707ae5b07712" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.188306 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerDied","Data":"a255b4fc4d71a17bfa31390573549daada5264701abd9f080f44707ae5b07712"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.188760 4926 scope.go:117] "RemoveContainer" containerID="a255b4fc4d71a17bfa31390573549daada5264701abd9f080f44707ae5b07712" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.190488 4926 generic.go:334] "Generic (PLEG): container finished" podID="409d9035-812c-4b37-b495-d6554315fb03" containerID="b7da2e7b4d7448d9ed118ae4e0a1e12664e5f491ef10105c0c16f7ef07497164" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.190556 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerDied","Data":"b7da2e7b4d7448d9ed118ae4e0a1e12664e5f491ef10105c0c16f7ef07497164"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.191099 4926 scope.go:117] "RemoveContainer" containerID="b7da2e7b4d7448d9ed118ae4e0a1e12664e5f491ef10105c0c16f7ef07497164" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.193274 4926 generic.go:334] "Generic (PLEG): container finished" podID="8292ff2d-08ba-4be8-a268-5c8b6d639087" containerID="49001810fcfbfa6ca05256a32bb07b914fa3e240a471bf9cd81c5af01a0420ba" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.193339 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" event={"ID":"8292ff2d-08ba-4be8-a268-5c8b6d639087","Type":"ContainerDied","Data":"49001810fcfbfa6ca05256a32bb07b914fa3e240a471bf9cd81c5af01a0420ba"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.193922 4926 scope.go:117] "RemoveContainer" containerID="49001810fcfbfa6ca05256a32bb07b914fa3e240a471bf9cd81c5af01a0420ba" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.196130 4926 generic.go:334] "Generic (PLEG): container finished" podID="194893a3-1c67-4a28-a67e-7d7eeef304a7" containerID="4a584e53b5952c702aa4f8c41c5f91bcf3aa30feb0caeb6e4699d6872ebb9580" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.196186 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" event={"ID":"194893a3-1c67-4a28-a67e-7d7eeef304a7","Type":"ContainerDied","Data":"4a584e53b5952c702aa4f8c41c5f91bcf3aa30feb0caeb6e4699d6872ebb9580"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.196515 4926 scope.go:117] "RemoveContainer" containerID="4a584e53b5952c702aa4f8c41c5f91bcf3aa30feb0caeb6e4699d6872ebb9580" Nov 25 10:22:37 crc kubenswrapper[4926]: E1125 10:22:37.196711 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-fr2mr_openstack-operators(194893a3-1c67-4a28-a67e-7d7eeef304a7)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" podUID="194893a3-1c67-4a28-a67e-7d7eeef304a7" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.197631 4926 generic.go:334] "Generic (PLEG): container finished" podID="dd863871-17e8-4a97-bc8b-5dd02f807225" containerID="06dfe4f5a2264dfe2ffe5ca14a1a5014464986fb1fbcd14a804f41cf46f4396e" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.197680 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerDied","Data":"06dfe4f5a2264dfe2ffe5ca14a1a5014464986fb1fbcd14a804f41cf46f4396e"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.198000 4926 scope.go:117] "RemoveContainer" containerID="06dfe4f5a2264dfe2ffe5ca14a1a5014464986fb1fbcd14a804f41cf46f4396e" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.200667 4926 generic.go:334] "Generic (PLEG): container finished" podID="0ed1bf71-b826-4854-873d-7c6fcf992bbd" containerID="a521e3270b8748427a6452a581a6d324467194c27ccf125143a4ae250a834174" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.200739 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerDied","Data":"a521e3270b8748427a6452a581a6d324467194c27ccf125143a4ae250a834174"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.201185 4926 scope.go:117] "RemoveContainer" containerID="a521e3270b8748427a6452a581a6d324467194c27ccf125143a4ae250a834174" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.206085 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b32eddd-fe58-4854-bc45-a97ec5349f86" containerID="d7b64d216e63148d402f7ada0737c596449e9f35674156f9098accb99a55c13f" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.206154 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerDied","Data":"d7b64d216e63148d402f7ada0737c596449e9f35674156f9098accb99a55c13f"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.206953 4926 scope.go:117] "RemoveContainer" containerID="d7b64d216e63148d402f7ada0737c596449e9f35674156f9098accb99a55c13f" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.208530 4926 generic.go:334] "Generic (PLEG): container finished" podID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" containerID="f74530828e1ab1cdd2254265df37ab63364c9bbe3016e486e11bf88a39d33a06" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.208591 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" event={"ID":"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e","Type":"ContainerDied","Data":"f74530828e1ab1cdd2254265df37ab63364c9bbe3016e486e11bf88a39d33a06"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.209007 4926 scope.go:117] "RemoveContainer" containerID="f74530828e1ab1cdd2254265df37ab63364c9bbe3016e486e11bf88a39d33a06" Nov 25 10:22:37 crc kubenswrapper[4926]: E1125 10:22:37.209248 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-864885998-vtrb2_openstack-operators(d95f6ac6-9ecf-4cef-ad4e-1389a10f303e)\"" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" podUID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.210169 4926 generic.go:334] "Generic (PLEG): container finished" podID="09b3e2a6-3249-4ca3-afc9-39dd744ddd10" containerID="ef61f457015a05047bb8f43a9915c81c83b01ac25deb26bff52245b232bf70be" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.210242 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerDied","Data":"ef61f457015a05047bb8f43a9915c81c83b01ac25deb26bff52245b232bf70be"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.210778 4926 scope.go:117] "RemoveContainer" containerID="ef61f457015a05047bb8f43a9915c81c83b01ac25deb26bff52245b232bf70be" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.213538 4926 scope.go:117] "RemoveContainer" containerID="413682c48013fdeed4fc2d5b9656cf12a302415171973b50a7bde547dee98ca6" Nov 25 10:22:37 crc kubenswrapper[4926]: E1125 10:22:37.213761 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-858778c9dc-lwgz6_openstack-operators(b6772bac-4625-41d7-915d-2fdabd9916b7)\"" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podUID="b6772bac-4625-41d7-915d-2fdabd9916b7" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.217742 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc152fdc-1bab-4528-992d-0883671d7372" containerID="2d3e6970086339beb02a13daa3467904448359265eaa86ed860edc6c8a7b8ff6" exitCode=1 Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.217852 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerDied","Data":"2d3e6970086339beb02a13daa3467904448359265eaa86ed860edc6c8a7b8ff6"} Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.217985 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.219350 4926 scope.go:117] "RemoveContainer" containerID="2d3e6970086339beb02a13daa3467904448359265eaa86ed860edc6c8a7b8ff6" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.610908 4926 scope.go:117] "RemoveContainer" containerID="cffdc6c8d780ff8826404234f8d89e8422500931f06a3fc4b92d62a5484033b9" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.786391 4926 scope.go:117] "RemoveContainer" containerID="9fcc3ce783d8ba3ae2c657c88c82388db47bbbb8acf55e583dcf67a89db85724" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.858101 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:22:37 crc kubenswrapper[4926]: I1125 10:22:37.862188 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.065397 4926 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" podUID="4403c995-1767-4968-b0f4-0a4f0d70a188" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.95:8081/readyz\": dial tcp 10.217.0.95:8081: connect: connection refused" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.232168 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerStarted","Data":"f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.238722 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerStarted","Data":"fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.241459 4926 generic.go:334] "Generic (PLEG): container finished" podID="4403c995-1767-4968-b0f4-0a4f0d70a188" containerID="ac3f82ab05b703ebd449519e2445a8ba3233383139cdd50b01dec1b6f0cf868b" exitCode=1 Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.241560 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" event={"ID":"4403c995-1767-4968-b0f4-0a4f0d70a188","Type":"ContainerDied","Data":"ac3f82ab05b703ebd449519e2445a8ba3233383139cdd50b01dec1b6f0cf868b"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.242349 4926 scope.go:117] "RemoveContainer" containerID="ac3f82ab05b703ebd449519e2445a8ba3233383139cdd50b01dec1b6f0cf868b" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.246131 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerStarted","Data":"7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.249797 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerStarted","Data":"4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.253968 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerStarted","Data":"dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.254182 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.258019 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerStarted","Data":"b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.261749 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerStarted","Data":"f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.262526 4926 scope.go:117] "RemoveContainer" containerID="f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0" Nov 25 10:22:38 crc kubenswrapper[4926]: E1125 10:22:38.262817 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-4d4g5_openstack-operators(df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.263802 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerStarted","Data":"5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.265420 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerStarted","Data":"e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.265589 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.268441 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerStarted","Data":"706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.272275 4926 generic.go:334] "Generic (PLEG): container finished" podID="f74db279-9026-4869-a147-d45bf581b618" containerID="b4ffe13bb865d6f9fe818fbe2b044605260bb23a2c50bb245bb0a4776a7eae26" exitCode=1 Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.272368 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" event={"ID":"f74db279-9026-4869-a147-d45bf581b618","Type":"ContainerDied","Data":"b4ffe13bb865d6f9fe818fbe2b044605260bb23a2c50bb245bb0a4776a7eae26"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.272434 4926 scope.go:117] "RemoveContainer" containerID="72594fd7a747457517e5f0c4214007b015ca36bca78226cee674fc1e7d3699f7" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.273175 4926 scope.go:117] "RemoveContainer" containerID="b4ffe13bb865d6f9fe818fbe2b044605260bb23a2c50bb245bb0a4776a7eae26" Nov 25 10:22:38 crc kubenswrapper[4926]: E1125 10:22:38.273430 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-mt4w9_openstack-operators(f74db279-9026-4869-a147-d45bf581b618)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" podUID="f74db279-9026-4869-a147-d45bf581b618" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.281394 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerID="f7956046f9d7cb816c6d7cf6ab39281c090e0f105f0b1f059972ec2e65667060" exitCode=1 Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.281471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerDied","Data":"f7956046f9d7cb816c6d7cf6ab39281c090e0f105f0b1f059972ec2e65667060"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.282147 4926 scope.go:117] "RemoveContainer" containerID="f7956046f9d7cb816c6d7cf6ab39281c090e0f105f0b1f059972ec2e65667060" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.285692 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerStarted","Data":"b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5"} Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.285897 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.303208 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 25 10:22:38 crc kubenswrapper[4926]: I1125 10:22:38.458561 4926 scope.go:117] "RemoveContainer" containerID="6fc501f19762f6105670f5e405746e249f6b955c4433daa84248acf85c9e864b" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.012223 4926 scope.go:117] "RemoveContainer" containerID="4d63c5b88246faa5057a3d692afb7f7d47cae3af0cdc6903a4a2a80ec084bbb7" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.306458 4926 generic.go:334] "Generic (PLEG): container finished" podID="5bf64213-939f-4f49-9030-9f9dcc78f8bb" containerID="fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.306708 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerDied","Data":"fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.306876 4926 scope.go:117] "RemoveContainer" containerID="3b6d22f86522965bb63d95715cfe147c87133d7900a3fc98424c22ab8736cb72" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.307236 4926 scope.go:117] "RemoveContainer" containerID="fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.307577 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-hgcnt_openstack-operators(5bf64213-939f-4f49-9030-9f9dcc78f8bb)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" podUID="5bf64213-939f-4f49-9030-9f9dcc78f8bb" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.315556 4926 generic.go:334] "Generic (PLEG): container finished" podID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" containerID="f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.315631 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerDied","Data":"f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.316780 4926 scope.go:117] "RemoveContainer" containerID="f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.317284 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-4d4g5_openstack-operators(df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.320973 4926 generic.go:334] "Generic (PLEG): container finished" podID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" containerID="2f06f1d35be777b814356442840720156176d5076b6d1f71ed272482bd4654a5" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.321031 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerDied","Data":"2f06f1d35be777b814356442840720156176d5076b6d1f71ed272482bd4654a5"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.322216 4926 scope.go:117] "RemoveContainer" containerID="2f06f1d35be777b814356442840720156176d5076b6d1f71ed272482bd4654a5" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.322770 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(2f8efeb7-cb75-4a17-bf87-85431d9e395e)\"" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.328231 4926 generic.go:334] "Generic (PLEG): container finished" podID="bc152fdc-1bab-4528-992d-0883671d7372" containerID="f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.328287 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerDied","Data":"f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.329041 4926 scope.go:117] "RemoveContainer" containerID="f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.329438 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-82vvs_openstack-operators(bc152fdc-1bab-4528-992d-0883671d7372)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" podUID="bc152fdc-1bab-4528-992d-0883671d7372" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.331029 4926 generic.go:334] "Generic (PLEG): container finished" podID="7b32eddd-fe58-4854-bc45-a97ec5349f86" containerID="7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.331083 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerDied","Data":"7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.331915 4926 scope.go:117] "RemoveContainer" containerID="7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.332453 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-5wdlm_openstack-operators(7b32eddd-fe58-4854-bc45-a97ec5349f86)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" podUID="7b32eddd-fe58-4854-bc45-a97ec5349f86" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.334425 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae1df4bc-d341-49fd-a49a-e02b4d8714d1" containerID="5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.334476 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerDied","Data":"5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.335108 4926 scope.go:117] "RemoveContainer" containerID="5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.335366 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-lwls8_openstack-operators(ae1df4bc-d341-49fd-a49a-e02b4d8714d1)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" podUID="ae1df4bc-d341-49fd-a49a-e02b4d8714d1" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.339504 4926 generic.go:334] "Generic (PLEG): container finished" podID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" containerID="f4d8beb63667c0427d57caf4abe235b6f043df7eefa1f4526cc43e51144c8da6" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.339572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" event={"ID":"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef","Type":"ContainerDied","Data":"f4d8beb63667c0427d57caf4abe235b6f043df7eefa1f4526cc43e51144c8da6"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.340283 4926 scope.go:117] "RemoveContainer" containerID="f4d8beb63667c0427d57caf4abe235b6f043df7eefa1f4526cc43e51144c8da6" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.340705 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-65f5fc5bb5-r6kqg_metallb-system(0cb1f5f8-6fcb-4664-8fcc-74e416d63eef)\"" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.344820 4926 generic.go:334] "Generic (PLEG): container finished" podID="4403c995-1767-4968-b0f4-0a4f0d70a188" containerID="eb3568bd769902ed035e1efbb6c653861beecc5173cc59cef32261ef3eda0b5c" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.344932 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" event={"ID":"4403c995-1767-4968-b0f4-0a4f0d70a188","Type":"ContainerDied","Data":"eb3568bd769902ed035e1efbb6c653861beecc5173cc59cef32261ef3eda0b5c"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.345897 4926 scope.go:117] "RemoveContainer" containerID="eb3568bd769902ed035e1efbb6c653861beecc5173cc59cef32261ef3eda0b5c" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.346244 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-86896d4c4b-ssjxq_openstack-operators(4403c995-1767-4968-b0f4-0a4f0d70a188)\"" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" podUID="4403c995-1767-4968-b0f4-0a4f0d70a188" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.349130 4926 generic.go:334] "Generic (PLEG): container finished" podID="dd863871-17e8-4a97-bc8b-5dd02f807225" containerID="b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.349194 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerDied","Data":"b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.350297 4926 scope.go:117] "RemoveContainer" containerID="b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.350637 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-947wd_openstack-operators(dd863871-17e8-4a97-bc8b-5dd02f807225)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" podUID="dd863871-17e8-4a97-bc8b-5dd02f807225" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.354061 4926 generic.go:334] "Generic (PLEG): container finished" podID="416161f0-e9fb-4909-83c0-dfe7ad15885f" containerID="6a7db3babadfc951e222ee22639f7d1772709e89ac00d07b67c1e6764b999a8a" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.354121 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" event={"ID":"416161f0-e9fb-4909-83c0-dfe7ad15885f","Type":"ContainerDied","Data":"6a7db3babadfc951e222ee22639f7d1772709e89ac00d07b67c1e6764b999a8a"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.354750 4926 scope.go:117] "RemoveContainer" containerID="6a7db3babadfc951e222ee22639f7d1772709e89ac00d07b67c1e6764b999a8a" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.355076 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-kspmf_openstack-operators(416161f0-e9fb-4909-83c0-dfe7ad15885f)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" podUID="416161f0-e9fb-4909-83c0-dfe7ad15885f" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.377966 4926 generic.go:334] "Generic (PLEG): container finished" podID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" containerID="4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.378127 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerDied","Data":"4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.378623 4926 scope.go:117] "RemoveContainer" containerID="4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.380490 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-75fd7f55fb-p7v2x_openstack-operators(5249ee3d-0121-42bf-bc4a-7d05a8410a9f)\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.381374 4926 generic.go:334] "Generic (PLEG): container finished" podID="09b3e2a6-3249-4ca3-afc9-39dd744ddd10" containerID="706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.381409 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerDied","Data":"706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.381979 4926 scope.go:117] "RemoveContainer" containerID="706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.382259 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-67llk_openstack-operators(09b3e2a6-3249-4ca3-afc9-39dd744ddd10)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" podUID="09b3e2a6-3249-4ca3-afc9-39dd744ddd10" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.383504 4926 scope.go:117] "RemoveContainer" containerID="a323101e1d43b04f7a693896c68109ac61b3f975edf283514ae070693da13635" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.383922 4926 generic.go:334] "Generic (PLEG): container finished" podID="409d9035-812c-4b37-b495-d6554315fb03" containerID="b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.384003 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerDied","Data":"b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.384854 4926 scope.go:117] "RemoveContainer" containerID="b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.385175 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-2b444_openstack-operators(409d9035-812c-4b37-b495-d6554315fb03)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.401360 4926 generic.go:334] "Generic (PLEG): container finished" podID="8292ff2d-08ba-4be8-a268-5c8b6d639087" containerID="0243f5a75f8f4a8c843b3222b9148af016d13235f25118abe6e7499de1f7012a" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.401436 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" event={"ID":"8292ff2d-08ba-4be8-a268-5c8b6d639087","Type":"ContainerDied","Data":"0243f5a75f8f4a8c843b3222b9148af016d13235f25118abe6e7499de1f7012a"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.402195 4926 scope.go:117] "RemoveContainer" containerID="0243f5a75f8f4a8c843b3222b9148af016d13235f25118abe6e7499de1f7012a" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.402530 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-4wxmj_openstack-operators(8292ff2d-08ba-4be8-a268-5c8b6d639087)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" podUID="8292ff2d-08ba-4be8-a268-5c8b6d639087" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.407511 4926 generic.go:334] "Generic (PLEG): container finished" podID="0ed1bf71-b826-4854-873d-7c6fcf992bbd" containerID="dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.407616 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerDied","Data":"dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.409011 4926 scope.go:117] "RemoveContainer" containerID="dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.409334 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-wt256_openstack-operators(0ed1bf71-b826-4854-873d-7c6fcf992bbd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" podUID="0ed1bf71-b826-4854-873d-7c6fcf992bbd" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.412188 4926 generic.go:334] "Generic (PLEG): container finished" podID="1154878f-e8a5-42de-916f-73276b979e74" containerID="e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4" exitCode=1 Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.413365 4926 scope.go:117] "RemoveContainer" containerID="e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4" Nov 25 10:22:39 crc kubenswrapper[4926]: E1125 10:22:39.413595 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-kqv6n_openstack-operators(1154878f-e8a5-42de-916f-73276b979e74)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" podUID="1154878f-e8a5-42de-916f-73276b979e74" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.413638 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerDied","Data":"e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4"} Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.444815 4926 scope.go:117] "RemoveContainer" containerID="f7956046f9d7cb816c6d7cf6ab39281c090e0f105f0b1f059972ec2e65667060" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.509203 4926 scope.go:117] "RemoveContainer" containerID="2d3e6970086339beb02a13daa3467904448359265eaa86ed860edc6c8a7b8ff6" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.576391 4926 scope.go:117] "RemoveContainer" containerID="d7b64d216e63148d402f7ada0737c596449e9f35674156f9098accb99a55c13f" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.630525 4926 scope.go:117] "RemoveContainer" containerID="8b6b76df23ca2d547afa40eacc6f3bce4d70e0b3f29f406496c7f0cb25b87a79" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.653027 4926 scope.go:117] "RemoveContainer" containerID="4d63c5b88246faa5057a3d692afb7f7d47cae3af0cdc6903a4a2a80ec084bbb7" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.685339 4926 scope.go:117] "RemoveContainer" containerID="ac3f82ab05b703ebd449519e2445a8ba3233383139cdd50b01dec1b6f0cf868b" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.719222 4926 scope.go:117] "RemoveContainer" containerID="06dfe4f5a2264dfe2ffe5ca14a1a5014464986fb1fbcd14a804f41cf46f4396e" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.742295 4926 scope.go:117] "RemoveContainer" containerID="fd3404684a77a8cab7d504c74b9a9f632755c7c1c63bff3fc75216a3e2b5c822" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.767285 4926 scope.go:117] "RemoveContainer" containerID="9dd1498849e62796c39436f05c633b26a66126e317de44db35d715b1d6a23a22" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.796856 4926 scope.go:117] "RemoveContainer" containerID="ef61f457015a05047bb8f43a9915c81c83b01ac25deb26bff52245b232bf70be" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.820629 4926 scope.go:117] "RemoveContainer" containerID="b7da2e7b4d7448d9ed118ae4e0a1e12664e5f491ef10105c0c16f7ef07497164" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.840738 4926 scope.go:117] "RemoveContainer" containerID="49001810fcfbfa6ca05256a32bb07b914fa3e240a471bf9cd81c5af01a0420ba" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.873056 4926 scope.go:117] "RemoveContainer" containerID="a521e3270b8748427a6452a581a6d324467194c27ccf125143a4ae250a834174" Nov 25 10:22:39 crc kubenswrapper[4926]: I1125 10:22:39.894139 4926 scope.go:117] "RemoveContainer" containerID="a255b4fc4d71a17bfa31390573549daada5264701abd9f080f44707ae5b07712" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.425987 4926 scope.go:117] "RemoveContainer" containerID="7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.426616 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-5wdlm_openstack-operators(7b32eddd-fe58-4854-bc45-a97ec5349f86)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" podUID="7b32eddd-fe58-4854-bc45-a97ec5349f86" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.429237 4926 scope.go:117] "RemoveContainer" containerID="5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.429518 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-lwls8_openstack-operators(ae1df4bc-d341-49fd-a49a-e02b4d8714d1)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" podUID="ae1df4bc-d341-49fd-a49a-e02b4d8714d1" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.438154 4926 scope.go:117] "RemoveContainer" containerID="dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.438457 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-wt256_openstack-operators(0ed1bf71-b826-4854-873d-7c6fcf992bbd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" podUID="0ed1bf71-b826-4854-873d-7c6fcf992bbd" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.439963 4926 scope.go:117] "RemoveContainer" containerID="2f06f1d35be777b814356442840720156176d5076b6d1f71ed272482bd4654a5" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.440221 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(2f8efeb7-cb75-4a17-bf87-85431d9e395e)\"" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.452907 4926 scope.go:117] "RemoveContainer" containerID="4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.453193 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-75fd7f55fb-p7v2x_openstack-operators(5249ee3d-0121-42bf-bc4a-7d05a8410a9f)\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.455224 4926 scope.go:117] "RemoveContainer" containerID="b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.455463 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-2b444_openstack-operators(409d9035-812c-4b37-b495-d6554315fb03)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.458836 4926 scope.go:117] "RemoveContainer" containerID="f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.459139 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-82vvs_openstack-operators(bc152fdc-1bab-4528-992d-0883671d7372)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" podUID="bc152fdc-1bab-4528-992d-0883671d7372" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.462784 4926 scope.go:117] "RemoveContainer" containerID="b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.463012 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-947wd_openstack-operators(dd863871-17e8-4a97-bc8b-5dd02f807225)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" podUID="dd863871-17e8-4a97-bc8b-5dd02f807225" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.467610 4926 scope.go:117] "RemoveContainer" containerID="e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.467875 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-kqv6n_openstack-operators(1154878f-e8a5-42de-916f-73276b979e74)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" podUID="1154878f-e8a5-42de-916f-73276b979e74" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.474567 4926 scope.go:117] "RemoveContainer" containerID="706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.475159 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-67llk_openstack-operators(09b3e2a6-3249-4ca3-afc9-39dd744ddd10)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" podUID="09b3e2a6-3249-4ca3-afc9-39dd744ddd10" Nov 25 10:22:40 crc kubenswrapper[4926]: I1125 10:22:40.480319 4926 scope.go:117] "RemoveContainer" containerID="fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef" Nov 25 10:22:40 crc kubenswrapper[4926]: E1125 10:22:40.480587 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-hgcnt_openstack-operators(5bf64213-939f-4f49-9030-9f9dcc78f8bb)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" podUID="5bf64213-939f-4f49-9030-9f9dcc78f8bb" Nov 25 10:22:42 crc kubenswrapper[4926]: I1125 10:22:42.573054 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 25 10:22:42 crc kubenswrapper[4926]: I1125 10:22:42.846469 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.379408 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.379775 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.380750 4926 scope.go:117] "RemoveContainer" containerID="4a584e53b5952c702aa4f8c41c5f91bcf3aa30feb0caeb6e4699d6872ebb9580" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.381189 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=barbican-operator-controller-manager-86dc4d89c8-fr2mr_openstack-operators(194893a3-1c67-4a28-a67e-7d7eeef304a7)\"" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" podUID="194893a3-1c67-4a28-a67e-7d7eeef304a7" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.414187 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.414246 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.415584 4926 scope.go:117] "RemoveContainer" containerID="706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.415920 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=cinder-operator-controller-manager-79856dc55c-67llk_openstack-operators(09b3e2a6-3249-4ca3-afc9-39dd744ddd10)\"" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" podUID="09b3e2a6-3249-4ca3-afc9-39dd744ddd10" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.440860 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.440904 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.441719 4926 scope.go:117] "RemoveContainer" containerID="0243f5a75f8f4a8c843b3222b9148af016d13235f25118abe6e7499de1f7012a" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.442172 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=designate-operator-controller-manager-7d695c9b56-4wxmj_openstack-operators(8292ff2d-08ba-4be8-a268-5c8b6d639087)\"" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" podUID="8292ff2d-08ba-4be8-a268-5c8b6d639087" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.447268 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.447378 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.448661 4926 scope.go:117] "RemoveContainer" containerID="4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.449001 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-75fd7f55fb-p7v2x_openstack-operators(5249ee3d-0121-42bf-bc4a-7d05a8410a9f)\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.506397 4926 scope.go:117] "RemoveContainer" containerID="4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.506675 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-75fd7f55fb-p7v2x_openstack-operators(5249ee3d-0121-42bf-bc4a-7d05a8410a9f)\"" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" podUID="5249ee3d-0121-42bf-bc4a-7d05a8410a9f" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.581118 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.581600 4926 scope.go:117] "RemoveContainer" containerID="e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.581914 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-68c9694994-kqv6n_openstack-operators(1154878f-e8a5-42de-916f-73276b979e74)\"" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" podUID="1154878f-e8a5-42de-916f-73276b979e74" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.594860 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.595664 4926 scope.go:117] "RemoveContainer" containerID="dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.595961 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ironic-operator-controller-manager-5bfcdc958c-wt256_openstack-operators(0ed1bf71-b826-4854-873d-7c6fcf992bbd)\"" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" podUID="0ed1bf71-b826-4854-873d-7c6fcf992bbd" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.748232 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.790721 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.790855 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.791477 4926 scope.go:117] "RemoveContainer" containerID="f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.791707 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-4d4g5_openstack-operators(df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.820911 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.820991 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.821616 4926 scope.go:117] "RemoveContainer" containerID="fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.821828 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-cb6c4fdb7-hgcnt_openstack-operators(5bf64213-939f-4f49-9030-9f9dcc78f8bb)\"" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" podUID="5bf64213-939f-4f49-9030-9f9dcc78f8bb" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.826171 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.827027 4926 scope.go:117] "RemoveContainer" containerID="b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.827258 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=neutron-operator-controller-manager-7c57c8bbc4-947wd_openstack-operators(dd863871-17e8-4a97-bc8b-5dd02f807225)\"" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" podUID="dd863871-17e8-4a97-bc8b-5dd02f807225" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.928844 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.928926 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.929987 4926 scope.go:117] "RemoveContainer" containerID="f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.930533 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-82vvs_openstack-operators(bc152fdc-1bab-4528-992d-0883671d7372)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" podUID="bc152fdc-1bab-4528-992d-0883671d7372" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.936053 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.936123 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.937012 4926 scope.go:117] "RemoveContainer" containerID="c8a513033d37f8bbe12f05de8260056f80f8b11aca124b6c67eb363a21f060c5" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.937303 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=nova-operator-controller-manager-79556f57fc-tmxx7_openstack-operators(fa24e0a8-f9a4-4160-bac5-4d73e8579fd8)\"" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" podUID="fa24e0a8-f9a4-4160-bac5-4d73e8579fd8" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.974071 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.974582 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:22:43 crc kubenswrapper[4926]: I1125 10:22:43.975289 4926 scope.go:117] "RemoveContainer" containerID="b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7" Nov 25 10:22:43 crc kubenswrapper[4926]: E1125 10:22:43.975533 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-2b444_openstack-operators(409d9035-812c-4b37-b495-d6554315fb03)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.014289 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.034689 4926 scope.go:117] "RemoveContainer" containerID="b4ffe13bb865d6f9fe818fbe2b044605260bb23a2c50bb245bb0a4776a7eae26" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.035131 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-mt4w9_openstack-operators(f74db279-9026-4869-a147-d45bf581b618)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" podUID="f74db279-9026-4869-a147-d45bf581b618" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.050219 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.111555 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.111596 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.112562 4926 scope.go:117] "RemoveContainer" containerID="c90c02ed94327d5dcc6994a7fa3e6e6d37115fbfc0e824c591c37fbae1eb196e" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.113019 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=ovn-operator-controller-manager-66cf5c67ff-hr2gx_openstack-operators(edb138df-3c28-4b7d-b285-2fe43094ead3)\"" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" podUID="edb138df-3c28-4b7d-b285-2fe43094ead3" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.123117 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.206827 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.206928 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.207914 4926 scope.go:117] "RemoveContainer" containerID="5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.208234 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=placement-operator-controller-manager-5db546f9d9-lwls8_openstack-operators(ae1df4bc-d341-49fd-a49a-e02b4d8714d1)\"" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" podUID="ae1df4bc-d341-49fd-a49a-e02b4d8714d1" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.209913 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-2m5cf" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.224603 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.225666 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.225821 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.226463 4926 scope.go:117] "RemoveContainer" containerID="6a7db3babadfc951e222ee22639f7d1772709e89ac00d07b67c1e6764b999a8a" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.226717 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-kspmf_openstack-operators(416161f0-e9fb-4909-83c0-dfe7ad15885f)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" podUID="416161f0-e9fb-4909-83c0-dfe7ad15885f" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.262503 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.262565 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.263674 4926 scope.go:117] "RemoveContainer" containerID="7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.263928 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=telemetry-operator-controller-manager-567f98c9d-5wdlm_openstack-operators(7b32eddd-fe58-4854-bc45-a97ec5349f86)\"" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" podUID="7b32eddd-fe58-4854-bc45-a97ec5349f86" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.317765 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.329502 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5cb74df96-fbbkn" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.340487 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.386272 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.386337 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.387362 4926 scope.go:117] "RemoveContainer" containerID="f74530828e1ab1cdd2254265df37ab63364c9bbe3016e486e11bf88a39d33a06" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.387688 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=watcher-operator-controller-manager-864885998-vtrb2_openstack-operators(d95f6ac6-9ecf-4cef-ad4e-1389a10f303e)\"" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" podUID="d95f6ac6-9ecf-4cef-ad4e-1389a10f303e" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.485314 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.514748 4926 scope.go:117] "RemoveContainer" containerID="f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.514999 4926 scope.go:117] "RemoveContainer" containerID="6a7db3babadfc951e222ee22639f7d1772709e89ac00d07b67c1e6764b999a8a" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.515051 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-748dc6576f-82vvs_openstack-operators(bc152fdc-1bab-4528-992d-0883671d7372)\"" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" podUID="bc152fdc-1bab-4528-992d-0883671d7372" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.515135 4926 scope.go:117] "RemoveContainer" containerID="b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.515243 4926 scope.go:117] "RemoveContainer" containerID="f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.515360 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=manila-operator-controller-manager-58bb8d67cc-2b444_openstack-operators(409d9035-812c-4b37-b495-d6554315fb03)\"" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" podUID="409d9035-812c-4b37-b495-d6554315fb03" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.515380 4926 scope.go:117] "RemoveContainer" containerID="b4ffe13bb865d6f9fe818fbe2b044605260bb23a2c50bb245bb0a4776a7eae26" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.515556 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=heat-operator-controller-manager-774b86978c-4d4g5_openstack-operators(df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a)\"" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" podUID="df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.515602 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=octavia-operator-controller-manager-fd75fd47d-mt4w9_openstack-operators(f74db279-9026-4869-a147-d45bf581b618)\"" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" podUID="f74db279-9026-4869-a147-d45bf581b618" Nov 25 10:22:44 crc kubenswrapper[4926]: E1125 10:22:44.515657 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-6fdc4fcf86-kspmf_openstack-operators(416161f0-e9fb-4909-83c0-dfe7ad15885f)\"" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" podUID="416161f0-e9fb-4909-83c0-dfe7ad15885f" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.649821 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.667156 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.711791 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.761425 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.769412 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.833199 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.856447 4926 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.874692 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.883045 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.896124 4926 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 25 10:22:44 crc kubenswrapper[4926]: I1125 10:22:44.957754 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.044890 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-krm2h" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.155046 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.359905 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.359989 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.360761 4926 scope.go:117] "RemoveContainer" containerID="413682c48013fdeed4fc2d5b9656cf12a302415171973b50a7bde547dee98ca6" Nov 25 10:22:45 crc kubenswrapper[4926]: E1125 10:22:45.361051 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-858778c9dc-lwgz6_openstack-operators(b6772bac-4625-41d7-915d-2fdabd9916b7)\"" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" podUID="b6772bac-4625-41d7-915d-2fdabd9916b7" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.472192 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.491087 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.491133 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.491823 4926 scope.go:117] "RemoveContainer" containerID="2f06f1d35be777b814356442840720156176d5076b6d1f71ed272482bd4654a5" Nov 25 10:22:45 crc kubenswrapper[4926]: E1125 10:22:45.492084 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-state-metrics\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-state-metrics pod=kube-state-metrics-0_openstack(2f8efeb7-cb75-4a17-bf87-85431d9e395e)\"" pod="openstack/kube-state-metrics-0" podUID="2f8efeb7-cb75-4a17-bf87-85431d9e395e" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.748734 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.765063 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.792203 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-dzd4c" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.865452 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.868051 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-rrrds" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.913732 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 25 10:22:45 crc kubenswrapper[4926]: I1125 10:22:45.917115 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.239906 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.243778 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.284816 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.293418 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.306113 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.318430 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-c77m5" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.436586 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.442443 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.558854 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.581735 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.734523 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.784489 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.846223 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 25 10:22:46 crc kubenswrapper[4926]: I1125 10:22:46.912431 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.023743 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.077519 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.103335 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.151616 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.319810 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-9rsbs" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.380374 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.408802 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-6444d758f4-lgw8g" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.457977 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.458585 4926 scope.go:117] "RemoveContainer" containerID="f4d8beb63667c0427d57caf4abe235b6f043df7eefa1f4526cc43e51144c8da6" Nov 25 10:22:47 crc kubenswrapper[4926]: E1125 10:22:47.458790 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-65f5fc5bb5-r6kqg_metallb-system(0cb1f5f8-6fcb-4664-8fcc-74e416d63eef)\"" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" podUID="0cb1f5f8-6fcb-4664-8fcc-74e416d63eef" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.559178 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.561592 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.691553 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.695221 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.704814 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.728146 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-2q4hl" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.811084 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-hrtpv" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.831202 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.847273 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.914604 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-llpch" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.933357 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.959546 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.967544 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-mgcqh" Nov 25 10:22:47 crc kubenswrapper[4926]: I1125 10:22:47.990730 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.057912 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.064088 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.064142 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.064914 4926 scope.go:117] "RemoveContainer" containerID="eb3568bd769902ed035e1efbb6c653861beecc5173cc59cef32261ef3eda0b5c" Nov 25 10:22:48 crc kubenswrapper[4926]: E1125 10:22:48.065331 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=openstack-operator-controller-manager-86896d4c4b-ssjxq_openstack-operators(4403c995-1767-4968-b0f4-0a4f0d70a188)\"" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" podUID="4403c995-1767-4968-b0f4-0a4f0d70a188" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.078514 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.103658 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.134212 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.171243 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.248069 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.362573 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.436641 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.555658 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.595625 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.599586 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.638692 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.660364 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.782347 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.791714 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.801181 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.805093 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.807132 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.812409 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.819719 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-txn9r" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.821680 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.851807 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.891034 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.900062 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pp2zc" Nov 25 10:22:48 crc kubenswrapper[4926]: I1125 10:22:48.947269 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-6g2nb" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.003429 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.012237 4926 scope.go:117] "RemoveContainer" containerID="2999aabfa41dbf288b61da882776f755069cdfa4ba9aac180b62ddc64f4b55ec" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.018961 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.034347 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.040205 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.046486 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.110109 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.133582 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.211750 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.224923 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.225657 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.261743 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.264536 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.305955 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.344965 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.346577 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rvl6x" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.349887 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.389869 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.423331 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.437500 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.482985 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.487077 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.595158 4926 generic.go:334] "Generic (PLEG): container finished" podID="03cc50ce-1146-4953-901e-19d5bd7c468e" containerID="0f269c5b476b81211dfd10360075ed0e5e82d346354e099c1945e5c04159d22d" exitCode=1 Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.595207 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" event={"ID":"03cc50ce-1146-4953-901e-19d5bd7c468e","Type":"ContainerDied","Data":"0f269c5b476b81211dfd10360075ed0e5e82d346354e099c1945e5c04159d22d"} Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.595243 4926 scope.go:117] "RemoveContainer" containerID="2999aabfa41dbf288b61da882776f755069cdfa4ba9aac180b62ddc64f4b55ec" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.595977 4926 scope.go:117] "RemoveContainer" containerID="0f269c5b476b81211dfd10360075ed0e5e82d346354e099c1945e5c04159d22d" Nov 25 10:22:49 crc kubenswrapper[4926]: E1125 10:22:49.596353 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-zdqj6_openstack-operators(03cc50ce-1146-4953-901e-19d5bd7c468e)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podUID="03cc50ce-1146-4953-901e-19d5bd7c468e" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.622799 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.623911 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.672774 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.672911 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.673843 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-bswhl" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.716961 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.766839 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.793846 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.797667 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.915130 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.915763 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.963148 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Nov 25 10:22:49 crc kubenswrapper[4926]: I1125 10:22:49.965108 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.063728 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.083387 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.095020 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.110783 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.130377 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.138402 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.183722 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.237908 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.257242 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6vlnp" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.273447 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.281466 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.299705 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.343232 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.359072 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.360516 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.378423 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.404240 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.455337 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.455896 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-qqs4j" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.482984 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-4plth" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.497235 4926 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.522745 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.635179 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.635418 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.640576 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.645295 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-76lks" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.708791 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.714626 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.721528 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.743508 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.798422 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.808441 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.845024 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 25 10:22:50 crc kubenswrapper[4926]: I1125 10:22:50.881949 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.084191 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-5ntjb" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.112306 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.117402 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.158631 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.158991 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.183956 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.233133 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.266090 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.298859 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.300587 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.320441 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.324704 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.346324 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.359016 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.360409 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.373427 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.374471 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-9dq2x" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.425302 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.433703 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.437218 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.467138 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.492609 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.507544 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.528694 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.545741 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.556652 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.559946 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.611127 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-8r2v8" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.620882 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.621416 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.716086 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.740001 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.793068 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.796256 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.814039 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.833417 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.870191 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.879480 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.898929 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 25 10:22:51 crc kubenswrapper[4926]: I1125 10:22:51.965164 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.011013 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.015455 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.018542 4926 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.032497 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-nzjdg" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.039020 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.058836 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.059278 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.074221 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-zmqcr" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.123295 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.144108 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.160352 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.215988 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.275933 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-2fvj4" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.327282 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.328611 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.335522 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-2c78d" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.391421 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-vvbg6" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.611142 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.614597 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.646102 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.693205 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-97ktp" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.702893 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.721778 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.781155 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.787341 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.929101 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.938464 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 25 10:22:52 crc kubenswrapper[4926]: I1125 10:22:52.945587 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.013831 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.015565 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-wxp4z" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.027089 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.032246 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.055662 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-cpxvl" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.080814 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.184071 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.198788 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.330090 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.337168 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.456836 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dkvwv" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.506721 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.563226 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-5jv27" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.568330 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.623554 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.684790 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.718361 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.767543 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.797525 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.846380 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.850332 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 25 10:22:53 crc kubenswrapper[4926]: I1125 10:22:53.882377 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.076281 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.104122 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.120063 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.138624 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.242212 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.262294 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.270009 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.275423 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.311013 4926 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.318869 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mx46t" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.357817 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.359908 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.367378 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.401201 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.409319 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.423563 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.428681 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.430359 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.459271 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.515822 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.519376 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.521294 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.564702 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.626336 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.691497 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.842495 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.843879 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-gb2hj" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.845092 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-bqb7m" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.845251 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.855924 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.904086 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.916342 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Nov 25 10:22:54 crc kubenswrapper[4926]: I1125 10:22:54.964602 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.006673 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.011920 4926 scope.go:117] "RemoveContainer" containerID="4a584e53b5952c702aa4f8c41c5f91bcf3aa30feb0caeb6e4699d6872ebb9580" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.012077 4926 scope.go:117] "RemoveContainer" containerID="f74530828e1ab1cdd2254265df37ab63364c9bbe3016e486e11bf88a39d33a06" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.012212 4926 scope.go:117] "RemoveContainer" containerID="c8a513033d37f8bbe12f05de8260056f80f8b11aca124b6c67eb363a21f060c5" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.012266 4926 scope.go:117] "RemoveContainer" containerID="b1377e9b801599f5ba9f2014c277ce45ec2ade2bba157b349a48e256603641b7" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.012378 4926 scope.go:117] "RemoveContainer" containerID="fe4b00207e0649730d0c2bd74d754f6b28e8f816acbfb73b04ad54418512e3ef" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.012766 4926 scope.go:117] "RemoveContainer" containerID="c90c02ed94327d5dcc6994a7fa3e6e6d37115fbfc0e824c591c37fbae1eb196e" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.013345 4926 scope.go:117] "RemoveContainer" containerID="f6de9482e5fe0efa2373e4daf135c1a50ad9853c37e354727c6fd13c231fc0f0" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.033445 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.054664 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.072601 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.155359 4926 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vmcx9" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.168829 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.176431 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.188145 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.211417 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.224662 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-vwm7h" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.279050 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.286099 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.348912 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-gsprt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.360231 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.391141 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.405095 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.412891 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.440757 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.522984 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.592643 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.620238 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-bhpdd" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.646931 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.647245 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" event={"ID":"edb138df-3c28-4b7d-b285-2fe43094ead3","Type":"ContainerStarted","Data":"93d07a72ac20700f236c2e6df92b5c49f26e63829b4cd759478aff2d6c0a37c5"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.647533 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.651580 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" event={"ID":"fa24e0a8-f9a4-4160-bac5-4d73e8579fd8","Type":"ContainerStarted","Data":"a9dc23b28488cc50c8678fa5b197de8f92ab6d96fdbe896fd61f83a50f13e592"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.651782 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.653894 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" event={"ID":"194893a3-1c67-4a28-a67e-7d7eeef304a7","Type":"ContainerStarted","Data":"d10ce69a191513ca73a2fe8002ca42732032c6cc270d2f197c25d2bebd0682ad"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.654132 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.656844 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" event={"ID":"409d9035-812c-4b37-b495-d6554315fb03","Type":"ContainerStarted","Data":"1e439b8cc59cbe0ffb6b57c6e6a0688ed4248577bebf393a29f417974887deda"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.657001 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.664076 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" event={"ID":"5bf64213-939f-4f49-9030-9f9dcc78f8bb","Type":"ContainerStarted","Data":"a5068a85f23c38ec0f84f99f591c06ecbc76fa51e29067175e7bf2f94e644872"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.665083 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.667919 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" event={"ID":"d95f6ac6-9ecf-4cef-ad4e-1389a10f303e","Type":"ContainerStarted","Data":"39ca671e6938fea10f51599c11af9bb9dd8bb0e284c9f59087cb335e27474b12"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.668738 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.670720 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" event={"ID":"df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a","Type":"ContainerStarted","Data":"220df45755ed9301f261cbbf91ef5ed99ac5bb5b51b34497e81d166e90bbe019"} Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.671209 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.675136 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.682442 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.708885 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.711469 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.711651 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.730364 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.774626 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.775311 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-h85ln" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.804670 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.807189 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.835564 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.956465 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-p77xz" Nov 25 10:22:55 crc kubenswrapper[4926]: I1125 10:22:55.992838 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.000338 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.002108 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.012718 4926 scope.go:117] "RemoveContainer" containerID="dbea497837e72c2911265a003925cc9649f44be43d67b0c0b2d8817365b6b8f7" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.012815 4926 scope.go:117] "RemoveContainer" containerID="706aa3f8ea98c96ecb9eebff7528c5da95bcf17b46d9be70d3538baa7bfa3951" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.012851 4926 scope.go:117] "RemoveContainer" containerID="7fda6c6c7d41e90358b4e5a19f435d72601e9098dc8fe241f57b53d3eedffbd0" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.026730 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.033726 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.042749 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.083722 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.130419 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.170058 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.195374 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.199322 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.202318 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.282369 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.294174 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.303658 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-vwftx" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.379055 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.421375 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.425272 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ddhxn" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.465231 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.572080 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.576847 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.582130 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.594236 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.625203 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-dklj7" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.644896 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.651931 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.667722 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.680669 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" event={"ID":"7b32eddd-fe58-4854-bc45-a97ec5349f86","Type":"ContainerStarted","Data":"7896af19110e703888777e1fee48ca8c7d3c0cb4008e29de4ea605ac13fb22d6"} Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.680873 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.683078 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" event={"ID":"0ed1bf71-b826-4854-873d-7c6fcf992bbd","Type":"ContainerStarted","Data":"19b20a90113c160c4c4d66a05595c621ea902f07d9a7b837b789f69c9676931f"} Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.683270 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.685103 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" event={"ID":"09b3e2a6-3249-4ca3-afc9-39dd744ddd10","Type":"ContainerStarted","Data":"6708fb13f6d9935dd0ee6e35a8e0be6171cbbea3064c5e4059e703f3b9304a6e"} Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.727553 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.849640 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-fc6gp" Nov 25 10:22:56 crc kubenswrapper[4926]: I1125 10:22:56.969705 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.012141 4926 scope.go:117] "RemoveContainer" containerID="2f06f1d35be777b814356442840720156176d5076b6d1f71ed272482bd4654a5" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.012300 4926 scope.go:117] "RemoveContainer" containerID="6a7db3babadfc951e222ee22639f7d1772709e89ac00d07b67c1e6764b999a8a" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.012447 4926 scope.go:117] "RemoveContainer" containerID="b4ffe13bb865d6f9fe818fbe2b044605260bb23a2c50bb245bb0a4776a7eae26" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.012899 4926 scope.go:117] "RemoveContainer" containerID="4f40507de5f097bdac09b95908fc386e771e501a2fa6064b0469a5b50a682531" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.079149 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.116862 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.130905 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.152892 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.174292 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bkprg" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.174511 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-gjcvm" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.181255 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.204328 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.340048 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-k8kl7" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.363513 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.373126 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.407492 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.417605 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.568125 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.639130 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.642194 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-cdkvl" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.652907 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.668314 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.671420 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.696340 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" event={"ID":"f74db279-9026-4869-a147-d45bf581b618","Type":"ContainerStarted","Data":"426cf24cf7846dbea6a1762295eba9c1d519de573a475f8a34d9990fc32f2c5d"} Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.696606 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.699759 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" event={"ID":"416161f0-e9fb-4909-83c0-dfe7ad15885f","Type":"ContainerStarted","Data":"2c630f17cc4352ce4b976995188a9cdecddc55db79275910f4d25c82f6076097"} Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.699977 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.701603 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" event={"ID":"5249ee3d-0121-42bf-bc4a-7d05a8410a9f","Type":"ContainerStarted","Data":"13c022a9459cbf1ab239180dc830bf3a44b563114582cb3b779a1b38e20aa2cc"} Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.701918 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.703890 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"2f8efeb7-cb75-4a17-bf87-85431d9e395e","Type":"ContainerStarted","Data":"cad2b835da0f0bed3f1fea4c049601636a801d1fd1e642756d14d002778c6028"} Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.704337 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.717592 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.718408 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-qvsws" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.745773 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.772791 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.773060 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.776101 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.839589 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-rr5v7" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.841622 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 25 10:22:57 crc kubenswrapper[4926]: I1125 10:22:57.950751 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.012497 4926 scope.go:117] "RemoveContainer" containerID="f6bc2606dbfce1b5cb3ee3d2b664c02d8fb0a1bf7121cdf1717de0766beba924" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.012901 4926 scope.go:117] "RemoveContainer" containerID="e8d40ef5bb8044bdc9ebbb0499b7f34f6e219f79f3686bac6e619fc36ea855e4" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.013030 4926 scope.go:117] "RemoveContainer" containerID="0243f5a75f8f4a8c843b3222b9148af016d13235f25118abe6e7499de1f7012a" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.096336 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.134205 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.205212 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.231252 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.231270 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.242776 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-r8hb9" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.391681 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.557553 4926 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.559066 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=46.559048768 podStartE2EDuration="46.559048768s" podCreationTimestamp="2025-11-25 10:22:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:22:32.572712673 +0000 UTC m=+2049.256669110" watchObservedRunningTime="2025-11-25 10:22:58.559048768 +0000 UTC m=+2075.243005195" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.564953 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.580221 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.580276 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wtrhn/must-gather-bzlp8","openshift-kube-apiserver/kube-apiserver-crc"] Nov 25 10:22:58 crc kubenswrapper[4926]: E1125 10:22:58.580655 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" containerName="installer" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.580672 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" containerName="installer" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.580875 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb9f2bfe-dd62-4c39-8864-f44219cfb999" containerName="installer" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.582282 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.584539 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wtrhn"/"kube-root-ca.crt" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.584708 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-wtrhn"/"openshift-service-ca.crt" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.587314 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.600617 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.603297 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=26.603277317 podStartE2EDuration="26.603277317s" podCreationTimestamp="2025-11-25 10:22:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:22:58.599447822 +0000 UTC m=+2075.283404239" watchObservedRunningTime="2025-11-25 10:22:58.603277317 +0000 UTC m=+2075.287233744" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.661461 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.680906 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-must-gather-output\") pod \"must-gather-bzlp8\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.681085 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccdm2\" (UniqueName: \"kubernetes.io/projected/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-kube-api-access-ccdm2\") pod \"must-gather-bzlp8\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.717388 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" event={"ID":"8292ff2d-08ba-4be8-a268-5c8b6d639087","Type":"ContainerStarted","Data":"40b064f2c3d80d6dfc1475427c6b739fe7485f3e96ad71c2ebdc3262be65d430"} Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.718065 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.720527 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" event={"ID":"1154878f-e8a5-42de-916f-73276b979e74","Type":"ContainerStarted","Data":"b2cb653cca752e164123dfa6328f96f6fe5360e0d9391185d6d68b3a467e57b1"} Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.721086 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.723743 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" event={"ID":"bc152fdc-1bab-4528-992d-0883671d7372","Type":"ContainerStarted","Data":"d2b2ef27d2c762f95f0eca277660eeb82e2fefc4b3998eda60a01235a2a73ce1"} Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.781691 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-must-gather-output\") pod \"must-gather-bzlp8\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.781821 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccdm2\" (UniqueName: \"kubernetes.io/projected/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-kube-api-access-ccdm2\") pod \"must-gather-bzlp8\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.782412 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-must-gather-output\") pod \"must-gather-bzlp8\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.786130 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.801724 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccdm2\" (UniqueName: \"kubernetes.io/projected/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-kube-api-access-ccdm2\") pod \"must-gather-bzlp8\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.834296 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.907752 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 25 10:22:58 crc kubenswrapper[4926]: I1125 10:22:58.908216 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.016757 4926 scope.go:117] "RemoveContainer" containerID="b92a9d1d5981d33b385c6935e42140861ef7d33db52fe2bf0db253d89b3c57a5" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.017141 4926 scope.go:117] "RemoveContainer" containerID="413682c48013fdeed4fc2d5b9656cf12a302415171973b50a7bde547dee98ca6" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.017484 4926 scope.go:117] "RemoveContainer" containerID="5c4eef073447ae927905213c050cc037f129067409b6d901778d5dd1e931ba0f" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.017834 4926 scope.go:117] "RemoveContainer" containerID="eb3568bd769902ed035e1efbb6c653861beecc5173cc59cef32261ef3eda0b5c" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.084434 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.157135 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.180968 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.193435 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.194256 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.230310 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.266786 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.367699 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-wtrhn/must-gather-bzlp8"] Nov 25 10:22:59 crc kubenswrapper[4926]: W1125 10:22:59.381890 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89fb69a6_8f3d_4f5d_b3f9_104be90c2d15.slice/crio-7f2a89d96349ee10ef025b649ad69beb4b363a52ba8abcad61f5d038718cc58e WatchSource:0}: Error finding container 7f2a89d96349ee10ef025b649ad69beb4b363a52ba8abcad61f5d038718cc58e: Status 404 returned error can't find the container with id 7f2a89d96349ee10ef025b649ad69beb4b363a52ba8abcad61f5d038718cc58e Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.414235 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.480993 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.720314 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.734695 4926 generic.go:334] "Generic (PLEG): container finished" podID="96a34dc4-fd0b-43c2-bed3-741482089f7b" containerID="401863c86ff6640ad7310a29bb44937ce36860976b9e7f5e2f82be773fc50970" exitCode=1 Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.734771 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" event={"ID":"96a34dc4-fd0b-43c2-bed3-741482089f7b","Type":"ContainerDied","Data":"401863c86ff6640ad7310a29bb44937ce36860976b9e7f5e2f82be773fc50970"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.735402 4926 scope.go:117] "RemoveContainer" containerID="401863c86ff6640ad7310a29bb44937ce36860976b9e7f5e2f82be773fc50970" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.739560 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" event={"ID":"dd863871-17e8-4a97-bc8b-5dd02f807225","Type":"ContainerStarted","Data":"ec0e404e42e76e99d1b06030154e2057c51855a6698996cb3b507620cfb5631f"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.739862 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.745677 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" event={"ID":"ae1df4bc-d341-49fd-a49a-e02b4d8714d1","Type":"ContainerStarted","Data":"d20fd4452de7161d9d6b96a077d494e4469eecf9adff7b7055110997353471ac"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.746167 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.747154 4926 generic.go:334] "Generic (PLEG): container finished" podID="1fddcfdd-0fb9-4153-a227-b28e6faccb40" containerID="cebb5b05bd3519f415c184b376d12a6d38734df3ea33fc71629be18ee5fb573e" exitCode=1 Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.747232 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" event={"ID":"1fddcfdd-0fb9-4153-a227-b28e6faccb40","Type":"ContainerDied","Data":"cebb5b05bd3519f415c184b376d12a6d38734df3ea33fc71629be18ee5fb573e"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.747592 4926 scope.go:117] "RemoveContainer" containerID="cebb5b05bd3519f415c184b376d12a6d38734df3ea33fc71629be18ee5fb573e" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.749579 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" event={"ID":"b6772bac-4625-41d7-915d-2fdabd9916b7","Type":"ContainerStarted","Data":"f289e56b4bcfbb704eb0563dabfb55473193eef5ad5fa7a67f9e3e58d6c245f6"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.749909 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.751250 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" event={"ID":"4403c995-1767-4968-b0f4-0a4f0d70a188","Type":"ContainerStarted","Data":"472c56fabf8447c538e2889ed76a3dffbbcfa63b8f7967b88a78ff93ee82c98c"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.751427 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.752638 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" event={"ID":"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15","Type":"ContainerStarted","Data":"7f2a89d96349ee10ef025b649ad69beb4b363a52ba8abcad61f5d038718cc58e"} Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.765696 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 25 10:22:59 crc kubenswrapper[4926]: I1125 10:22:59.879043 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.340864 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.341068 4926 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.341172 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.341785 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.348023 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.348250 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.497298 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.504594 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.516615 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.633239 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.685786 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.706502 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.774088 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-hpkft" event={"ID":"96a34dc4-fd0b-43c2-bed3-741482089f7b","Type":"ContainerStarted","Data":"603487e9b66f0fe70b6a4ceb1e5ca86c8db4e2ec1ba430fda2af8b0c56452a88"} Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.776913 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-sxxx6" event={"ID":"1fddcfdd-0fb9-4153-a227-b28e6faccb40","Type":"ContainerStarted","Data":"79ffd60f46468598b78cf2099b526972c6bb29612d11348ace869884a88abd3d"} Nov 25 10:23:00 crc kubenswrapper[4926]: I1125 10:23:00.800395 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 25 10:23:01 crc kubenswrapper[4926]: I1125 10:23:01.012863 4926 scope.go:117] "RemoveContainer" containerID="0f269c5b476b81211dfd10360075ed0e5e82d346354e099c1945e5c04159d22d" Nov 25 10:23:01 crc kubenswrapper[4926]: E1125 10:23:01.013327 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-manager-668c99d594-zdqj6_openstack-operators(03cc50ce-1146-4953-901e-19d5bd7c468e)\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" podUID="03cc50ce-1146-4953-901e-19d5bd7c468e" Nov 25 10:23:01 crc kubenswrapper[4926]: I1125 10:23:01.116374 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 25 10:23:01 crc kubenswrapper[4926]: I1125 10:23:01.440012 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-lv9rl" Nov 25 10:23:01 crc kubenswrapper[4926]: I1125 10:23:01.594836 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 25 10:23:02 crc kubenswrapper[4926]: I1125 10:23:02.014710 4926 scope.go:117] "RemoveContainer" containerID="f4d8beb63667c0427d57caf4abe235b6f043df7eefa1f4526cc43e51144c8da6" Nov 25 10:23:02 crc kubenswrapper[4926]: I1125 10:23:02.419929 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Nov 25 10:23:02 crc kubenswrapper[4926]: I1125 10:23:02.616791 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 25 10:23:02 crc kubenswrapper[4926]: I1125 10:23:02.802137 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" event={"ID":"0cb1f5f8-6fcb-4664-8fcc-74e416d63eef","Type":"ContainerStarted","Data":"f3f21ece30cf882a0bbc770c1b2631c4d03f0a793d4246eb914a9efdda0668da"} Nov 25 10:23:02 crc kubenswrapper[4926]: I1125 10:23:02.802383 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.382363 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-86dc4d89c8-fr2mr" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.414332 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.421064 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-79856dc55c-67llk" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.448659 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-7d695c9b56-4wxmj" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.451411 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-75fd7f55fb-p7v2x" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.585576 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c9694994-kqv6n" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.596581 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-5bfcdc958c-wt256" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.793533 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-774b86978c-4d4g5" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.825391 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-cb6c4fdb7-hgcnt" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.937024 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.940785 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-79556f57fc-tmxx7" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.943593 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-748dc6576f-82vvs" Nov 25 10:23:03 crc kubenswrapper[4926]: I1125 10:23:03.981300 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-58bb8d67cc-2b444" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.056277 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-fd75fd47d-mt4w9" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.117329 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-66cf5c67ff-hr2gx" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.210120 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5db546f9d9-lwls8" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.231370 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-6fdc4fcf86-kspmf" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.267487 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-567f98c9d-5wdlm" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.390168 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-864885998-vtrb2" Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.839050 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" event={"ID":"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15","Type":"ContainerStarted","Data":"5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602"} Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.839358 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" event={"ID":"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15","Type":"ContainerStarted","Data":"585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22"} Nov 25 10:23:04 crc kubenswrapper[4926]: I1125 10:23:04.856103 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" podStartSLOduration=5.943353597 podStartE2EDuration="10.8560829s" podCreationTimestamp="2025-11-25 10:22:54 +0000 UTC" firstStartedPulling="2025-11-25 10:22:59.388286704 +0000 UTC m=+2076.072243141" lastFinishedPulling="2025-11-25 10:23:04.301016017 +0000 UTC m=+2080.984972444" observedRunningTime="2025-11-25 10:23:04.850778108 +0000 UTC m=+2081.534734555" watchObservedRunningTime="2025-11-25 10:23:04.8560829 +0000 UTC m=+2081.540039327" Nov 25 10:23:05 crc kubenswrapper[4926]: I1125 10:23:05.366470 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-858778c9dc-lwgz6" Nov 25 10:23:05 crc kubenswrapper[4926]: I1125 10:23:05.497572 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Nov 25 10:23:06 crc kubenswrapper[4926]: I1125 10:23:06.534318 4926 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:23:06 crc kubenswrapper[4926]: I1125 10:23:06.534801 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f" gracePeriod=5 Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.821952 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wtrhn/crc-debug-zw6x8"] Nov 25 10:23:07 crc kubenswrapper[4926]: E1125 10:23:07.822682 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.822696 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.822902 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.823500 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.893715 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/097dab84-8b0b-4bbc-af0f-18b062a1a891-host\") pod \"crc-debug-zw6x8\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.893777 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlcqf\" (UniqueName: \"kubernetes.io/projected/097dab84-8b0b-4bbc-af0f-18b062a1a891-kube-api-access-jlcqf\") pod \"crc-debug-zw6x8\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.995706 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/097dab84-8b0b-4bbc-af0f-18b062a1a891-host\") pod \"crc-debug-zw6x8\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.995766 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlcqf\" (UniqueName: \"kubernetes.io/projected/097dab84-8b0b-4bbc-af0f-18b062a1a891-kube-api-access-jlcqf\") pod \"crc-debug-zw6x8\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:07 crc kubenswrapper[4926]: I1125 10:23:07.995919 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/097dab84-8b0b-4bbc-af0f-18b062a1a891-host\") pod \"crc-debug-zw6x8\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:08 crc kubenswrapper[4926]: I1125 10:23:08.017743 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlcqf\" (UniqueName: \"kubernetes.io/projected/097dab84-8b0b-4bbc-af0f-18b062a1a891-kube-api-access-jlcqf\") pod \"crc-debug-zw6x8\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:08 crc kubenswrapper[4926]: I1125 10:23:08.071050 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-86896d4c4b-ssjxq" Nov 25 10:23:08 crc kubenswrapper[4926]: I1125 10:23:08.159371 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:08 crc kubenswrapper[4926]: W1125 10:23:08.228584 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod097dab84_8b0b_4bbc_af0f_18b062a1a891.slice/crio-513e6d3126260d4dfee5cbc9837716a16f83890e376f86647a08b4e14688032b WatchSource:0}: Error finding container 513e6d3126260d4dfee5cbc9837716a16f83890e376f86647a08b4e14688032b: Status 404 returned error can't find the container with id 513e6d3126260d4dfee5cbc9837716a16f83890e376f86647a08b4e14688032b Nov 25 10:23:08 crc kubenswrapper[4926]: I1125 10:23:08.871303 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" event={"ID":"097dab84-8b0b-4bbc-af0f-18b062a1a891","Type":"ContainerStarted","Data":"513e6d3126260d4dfee5cbc9837716a16f83890e376f86647a08b4e14688032b"} Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.672142 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.672426 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872417 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872514 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872543 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872576 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872631 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872782 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872816 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.872904 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.873072 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.873165 4926 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.873187 4926 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.873198 4926 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.886128 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.905931 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.906281 4926 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f" exitCode=137 Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.906367 4926 scope.go:117] "RemoveContainer" containerID="14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.906486 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.974308 4926 scope.go:117] "RemoveContainer" containerID="14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f" Nov 25 10:23:11 crc kubenswrapper[4926]: E1125 10:23:11.974957 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f\": container with ID starting with 14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f not found: ID does not exist" containerID="14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.975022 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f"} err="failed to get container status \"14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f\": rpc error: code = NotFound desc = could not find container \"14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f\": container with ID starting with 14672c86254dadc5f6c93c8bd4649715fc1a28d879fac31db1033b4bc84f235f not found: ID does not exist" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.975553 4926 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:11 crc kubenswrapper[4926]: I1125 10:23:11.975579 4926 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:12 crc kubenswrapper[4926]: I1125 10:23:12.032304 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 25 10:23:12 crc kubenswrapper[4926]: I1125 10:23:12.032710 4926 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Nov 25 10:23:12 crc kubenswrapper[4926]: I1125 10:23:12.050638 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:23:12 crc kubenswrapper[4926]: I1125 10:23:12.050686 4926 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="6ff49b17-11fa-45df-a486-b3be8f79663c" Nov 25 10:23:12 crc kubenswrapper[4926]: I1125 10:23:12.068025 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 25 10:23:12 crc kubenswrapper[4926]: I1125 10:23:12.068062 4926 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="6ff49b17-11fa-45df-a486-b3be8f79663c" Nov 25 10:23:13 crc kubenswrapper[4926]: I1125 10:23:13.829055 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-7c57c8bbc4-947wd" Nov 25 10:23:13 crc kubenswrapper[4926]: I1125 10:23:13.936565 4926 generic.go:334] "Generic (PLEG): container finished" podID="8e239ba8-a966-470d-b462-c63630f38b7f" containerID="60e71f9e5eeccec9e5fd287702d4d8b475309c3b1d255aa09006d2cf42ed7c65" exitCode=0 Nov 25 10:23:13 crc kubenswrapper[4926]: I1125 10:23:13.936814 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" event={"ID":"8e239ba8-a966-470d-b462-c63630f38b7f","Type":"ContainerDied","Data":"60e71f9e5eeccec9e5fd287702d4d8b475309c3b1d255aa09006d2cf42ed7c65"} Nov 25 10:23:13 crc kubenswrapper[4926]: I1125 10:23:13.937606 4926 scope.go:117] "RemoveContainer" containerID="60e71f9e5eeccec9e5fd287702d4d8b475309c3b1d255aa09006d2cf42ed7c65" Nov 25 10:23:16 crc kubenswrapper[4926]: I1125 10:23:16.013206 4926 scope.go:117] "RemoveContainer" containerID="0f269c5b476b81211dfd10360075ed0e5e82d346354e099c1945e5c04159d22d" Nov 25 10:23:19 crc kubenswrapper[4926]: I1125 10:23:19.996068 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" event={"ID":"8e239ba8-a966-470d-b462-c63630f38b7f","Type":"ContainerStarted","Data":"9b10e5b9ec40d96c06440027e6690f2bc0aed2f3469a400abe145a228a091302"} Nov 25 10:23:19 crc kubenswrapper[4926]: I1125 10:23:19.996981 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 10:23:19 crc kubenswrapper[4926]: I1125 10:23:19.999123 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-tkd8b" Nov 25 10:23:19 crc kubenswrapper[4926]: I1125 10:23:19.999881 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-zdqj6" event={"ID":"03cc50ce-1146-4953-901e-19d5bd7c468e","Type":"ContainerStarted","Data":"274d819a3ce97ae353cc962eef437df43951b7f4459a12898db251a0cd17bfc8"} Nov 25 10:23:21 crc kubenswrapper[4926]: I1125 10:23:21.023086 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" event={"ID":"097dab84-8b0b-4bbc-af0f-18b062a1a891","Type":"ContainerStarted","Data":"a030c75c8c794bfe9a30303013f9354277d780fa603207a12ab49243501550a3"} Nov 25 10:23:21 crc kubenswrapper[4926]: I1125 10:23:21.047439 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" podStartSLOduration=1.9666720020000001 podStartE2EDuration="14.047418273s" podCreationTimestamp="2025-11-25 10:23:07 +0000 UTC" firstStartedPulling="2025-11-25 10:23:08.232095779 +0000 UTC m=+2084.916052206" lastFinishedPulling="2025-11-25 10:23:20.31284206 +0000 UTC m=+2096.996798477" observedRunningTime="2025-11-25 10:23:21.037032885 +0000 UTC m=+2097.720989322" watchObservedRunningTime="2025-11-25 10:23:21.047418273 +0000 UTC m=+2097.731374700" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.766158 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vzjll"] Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.768899 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.788662 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vzjll"] Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.800278 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-catalog-content\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.800333 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fg4g6\" (UniqueName: \"kubernetes.io/projected/d043df18-3a9d-4803-b6b5-398a6e0a578b-kube-api-access-fg4g6\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.800369 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-utilities\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.901540 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-catalog-content\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.901882 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fg4g6\" (UniqueName: \"kubernetes.io/projected/d043df18-3a9d-4803-b6b5-398a6e0a578b-kube-api-access-fg4g6\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.901927 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-utilities\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.902121 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-catalog-content\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.902354 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-utilities\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:23 crc kubenswrapper[4926]: I1125 10:23:23.923283 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fg4g6\" (UniqueName: \"kubernetes.io/projected/d043df18-3a9d-4803-b6b5-398a6e0a578b-kube-api-access-fg4g6\") pod \"community-operators-vzjll\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:24 crc kubenswrapper[4926]: I1125 10:23:24.098559 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:24 crc kubenswrapper[4926]: I1125 10:23:24.575075 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vzjll"] Nov 25 10:23:25 crc kubenswrapper[4926]: I1125 10:23:25.060147 4926 generic.go:334] "Generic (PLEG): container finished" podID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerID="0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e" exitCode=0 Nov 25 10:23:25 crc kubenswrapper[4926]: I1125 10:23:25.060424 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerDied","Data":"0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e"} Nov 25 10:23:25 crc kubenswrapper[4926]: I1125 10:23:25.060450 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerStarted","Data":"c8a7f5768a619d295478b8d50df6dd85290efab104dc974e042b1ecc3292728b"} Nov 25 10:23:26 crc kubenswrapper[4926]: I1125 10:23:26.071557 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerStarted","Data":"43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff"} Nov 25 10:23:27 crc kubenswrapper[4926]: I1125 10:23:27.081383 4926 generic.go:334] "Generic (PLEG): container finished" podID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerID="43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff" exitCode=0 Nov 25 10:23:27 crc kubenswrapper[4926]: I1125 10:23:27.081441 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerDied","Data":"43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff"} Nov 25 10:23:28 crc kubenswrapper[4926]: I1125 10:23:28.092904 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerStarted","Data":"bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b"} Nov 25 10:23:28 crc kubenswrapper[4926]: I1125 10:23:28.123696 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vzjll" podStartSLOduration=2.6778345 podStartE2EDuration="5.123677697s" podCreationTimestamp="2025-11-25 10:23:23 +0000 UTC" firstStartedPulling="2025-11-25 10:23:25.067111935 +0000 UTC m=+2101.751068362" lastFinishedPulling="2025-11-25 10:23:27.512955122 +0000 UTC m=+2104.196911559" observedRunningTime="2025-11-25 10:23:28.115472044 +0000 UTC m=+2104.799428471" watchObservedRunningTime="2025-11-25 10:23:28.123677697 +0000 UTC m=+2104.807634124" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.366460 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tcmdr"] Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.369187 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.387456 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tcmdr"] Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.535274 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-utilities\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.535564 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-catalog-content\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.535623 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnvvs\" (UniqueName: \"kubernetes.io/projected/55cf4e17-622b-4238-894c-7b3362839762-kube-api-access-jnvvs\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.637625 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnvvs\" (UniqueName: \"kubernetes.io/projected/55cf4e17-622b-4238-894c-7b3362839762-kube-api-access-jnvvs\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.637813 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-utilities\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.637861 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-catalog-content\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.638499 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-catalog-content\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.639100 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-utilities\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.664579 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnvvs\" (UniqueName: \"kubernetes.io/projected/55cf4e17-622b-4238-894c-7b3362839762-kube-api-access-jnvvs\") pod \"community-operators-tcmdr\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:30 crc kubenswrapper[4926]: I1125 10:23:30.694432 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.361991 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6phfs"] Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.364814 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.382701 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6phfs"] Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.488077 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgf7w\" (UniqueName: \"kubernetes.io/projected/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-kube-api-access-rgf7w\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.488180 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-utilities\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.488221 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-catalog-content\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.590280 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-catalog-content\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.590465 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgf7w\" (UniqueName: \"kubernetes.io/projected/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-kube-api-access-rgf7w\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.590542 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-utilities\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.591036 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-catalog-content\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.591091 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-utilities\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.615864 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgf7w\" (UniqueName: \"kubernetes.io/projected/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-kube-api-access-rgf7w\") pod \"community-operators-6phfs\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:33 crc kubenswrapper[4926]: I1125 10:23:33.731483 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.099137 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.099181 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.144514 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.195841 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.767773 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jscf7"] Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.769864 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.776498 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jscf7"] Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.913756 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2s2c\" (UniqueName: \"kubernetes.io/projected/705621a2-d616-487e-b80a-187d8d6214c2-kube-api-access-l2s2c\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.913821 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-catalog-content\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:34 crc kubenswrapper[4926]: I1125 10:23:34.913984 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-utilities\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.015671 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-utilities\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.015796 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2s2c\" (UniqueName: \"kubernetes.io/projected/705621a2-d616-487e-b80a-187d8d6214c2-kube-api-access-l2s2c\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.015819 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-catalog-content\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.016175 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-utilities\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.016331 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-catalog-content\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.043998 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2s2c\" (UniqueName: \"kubernetes.io/projected/705621a2-d616-487e-b80a-187d8d6214c2-kube-api-access-l2s2c\") pod \"community-operators-jscf7\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.105351 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.985413 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-527cz"] Nov 25 10:23:35 crc kubenswrapper[4926]: I1125 10:23:35.989541 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.004594 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-527cz"] Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.136587 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-catalog-content\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.136806 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-utilities\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.137144 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sjv2\" (UniqueName: \"kubernetes.io/projected/a2275402-9d74-4709-9908-add9b0b816c4-kube-api-access-9sjv2\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.239054 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-utilities\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.239441 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sjv2\" (UniqueName: \"kubernetes.io/projected/a2275402-9d74-4709-9908-add9b0b816c4-kube-api-access-9sjv2\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.239559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-catalog-content\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.239916 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-catalog-content\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.240120 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-utilities\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.258428 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sjv2\" (UniqueName: \"kubernetes.io/projected/a2275402-9d74-4709-9908-add9b0b816c4-kube-api-access-9sjv2\") pod \"community-operators-527cz\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:36 crc kubenswrapper[4926]: I1125 10:23:36.308989 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.161213 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jdm9g"] Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.164108 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.177507 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jdm9g"] Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.257172 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-catalog-content\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.257301 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-utilities\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.257402 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86x69\" (UniqueName: \"kubernetes.io/projected/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-kube-api-access-86x69\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.359143 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-utilities\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.359486 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86x69\" (UniqueName: \"kubernetes.io/projected/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-kube-api-access-86x69\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.359559 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-catalog-content\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.360050 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-catalog-content\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.360320 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-utilities\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.387037 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86x69\" (UniqueName: \"kubernetes.io/projected/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-kube-api-access-86x69\") pod \"community-operators-jdm9g\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.461597 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-65f5fc5bb5-r6kqg" Nov 25 10:23:37 crc kubenswrapper[4926]: I1125 10:23:37.508761 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.408557 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4wgsp"] Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.422045 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.438054 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4wgsp"] Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.602568 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-utilities\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.602634 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7294z\" (UniqueName: \"kubernetes.io/projected/25b0e9bd-d961-4465-9496-14420a9daf2c-kube-api-access-7294z\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.602760 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-catalog-content\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.704013 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-utilities\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.704351 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7294z\" (UniqueName: \"kubernetes.io/projected/25b0e9bd-d961-4465-9496-14420a9daf2c-kube-api-access-7294z\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.704443 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-catalog-content\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.704771 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-utilities\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.704810 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-catalog-content\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.725469 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6phfs"] Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.736137 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7294z\" (UniqueName: \"kubernetes.io/projected/25b0e9bd-d961-4465-9496-14420a9daf2c-kube-api-access-7294z\") pod \"community-operators-4wgsp\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.762052 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.822153 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jscf7"] Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.869007 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tcmdr"] Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.913548 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jdm9g"] Nov 25 10:23:38 crc kubenswrapper[4926]: I1125 10:23:38.946982 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-527cz"] Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.221095 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jscf7" event={"ID":"705621a2-d616-487e-b80a-187d8d6214c2","Type":"ContainerStarted","Data":"eb4675e24ac0a438940812bce1372c24f9f23f0f2f84dcbca1df28fd61ff6c6b"} Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.223597 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-527cz" event={"ID":"a2275402-9d74-4709-9908-add9b0b816c4","Type":"ContainerStarted","Data":"cf9b8bbd48302c671c01616fe3c04880ac9cea82a51014a974a0636a2df79402"} Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.232469 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerStarted","Data":"410e46ce45266faa35eadca767ec1291556274bee9f84a8ccde45d0d027f4f79"} Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.239305 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerStarted","Data":"80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585"} Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.239360 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerStarted","Data":"3e65fac444114c7cf63e5eb1176a63807ace82c6083fb60d6e2d909e6d56409c"} Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.241240 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerStarted","Data":"0361a102815644dbdd61a2ab70d92d12bb42d01475df714a3cf494defb923d1f"} Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.385870 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4wgsp"] Nov 25 10:23:39 crc kubenswrapper[4926]: W1125 10:23:39.413502 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25b0e9bd_d961_4465_9496_14420a9daf2c.slice/crio-b1cd4a6da2d7b2f769e58f9825bbec5b222398045b7e2550804a3c001fb013a7 WatchSource:0}: Error finding container b1cd4a6da2d7b2f769e58f9825bbec5b222398045b7e2550804a3c001fb013a7: Status 404 returned error can't find the container with id b1cd4a6da2d7b2f769e58f9825bbec5b222398045b7e2550804a3c001fb013a7 Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.564745 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-stqkg"] Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.568332 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.575620 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stqkg"] Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.731028 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-utilities\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.731105 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-catalog-content\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.731132 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q7gv\" (UniqueName: \"kubernetes.io/projected/675b802f-5459-4178-a8b1-ad249451a47b-kube-api-access-5q7gv\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.832998 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-utilities\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.833071 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-catalog-content\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.833104 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q7gv\" (UniqueName: \"kubernetes.io/projected/675b802f-5459-4178-a8b1-ad249451a47b-kube-api-access-5q7gv\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.833594 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-utilities\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.833623 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-catalog-content\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.853744 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q7gv\" (UniqueName: \"kubernetes.io/projected/675b802f-5459-4178-a8b1-ad249451a47b-kube-api-access-5q7gv\") pod \"community-operators-stqkg\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:39 crc kubenswrapper[4926]: I1125 10:23:39.956060 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.255665 4926 generic.go:334] "Generic (PLEG): container finished" podID="705621a2-d616-487e-b80a-187d8d6214c2" containerID="91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385" exitCode=0 Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.255747 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jscf7" event={"ID":"705621a2-d616-487e-b80a-187d8d6214c2","Type":"ContainerDied","Data":"91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.270243 4926 generic.go:334] "Generic (PLEG): container finished" podID="a2275402-9d74-4709-9908-add9b0b816c4" containerID="e91972f16c3ba7fe0aff40b29213e8cfb5e9306b13a6d2501175fb712636ac84" exitCode=0 Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.270355 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-527cz" event={"ID":"a2275402-9d74-4709-9908-add9b0b816c4","Type":"ContainerDied","Data":"e91972f16c3ba7fe0aff40b29213e8cfb5e9306b13a6d2501175fb712636ac84"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.280059 4926 generic.go:334] "Generic (PLEG): container finished" podID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerID="4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53" exitCode=0 Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.280162 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerDied","Data":"4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.280217 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerStarted","Data":"b1cd4a6da2d7b2f769e58f9825bbec5b222398045b7e2550804a3c001fb013a7"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.290567 4926 generic.go:334] "Generic (PLEG): container finished" podID="55cf4e17-622b-4238-894c-7b3362839762" containerID="5aa5da072b249b90372df24f403b9372abe3d2aead066774905eea565d65b663" exitCode=0 Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.290672 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerDied","Data":"5aa5da072b249b90372df24f403b9372abe3d2aead066774905eea565d65b663"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.297160 4926 generic.go:334] "Generic (PLEG): container finished" podID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerID="80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585" exitCode=0 Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.297252 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerDied","Data":"80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.304204 4926 generic.go:334] "Generic (PLEG): container finished" podID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerID="1429a5922b3a590f2bf1144f8f98f3927e78b24d246ab9a712a81deb009e9a83" exitCode=0 Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.304253 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerDied","Data":"1429a5922b3a590f2bf1144f8f98f3927e78b24d246ab9a712a81deb009e9a83"} Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.456504 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-stqkg"] Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.766851 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4gpsl"] Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.769763 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.787359 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4gpsl"] Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.956544 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnrjj\" (UniqueName: \"kubernetes.io/projected/eab59112-b641-43af-a91b-e7c0e58ff417-kube-api-access-dnrjj\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.956902 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-catalog-content\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:40 crc kubenswrapper[4926]: I1125 10:23:40.956994 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-utilities\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.059168 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-catalog-content\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.059523 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-utilities\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.059685 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnrjj\" (UniqueName: \"kubernetes.io/projected/eab59112-b641-43af-a91b-e7c0e58ff417-kube-api-access-dnrjj\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.059911 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-utilities\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.060135 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-catalog-content\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.110730 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnrjj\" (UniqueName: \"kubernetes.io/projected/eab59112-b641-43af-a91b-e7c0e58ff417-kube-api-access-dnrjj\") pod \"community-operators-4gpsl\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.137990 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.327706 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerStarted","Data":"42a8de03ebbfdbb0aeaaa9cf0d2568fc81055cae01217b924427459ee714959e"} Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.337002 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerStarted","Data":"11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2"} Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.354604 4926 generic.go:334] "Generic (PLEG): container finished" podID="675b802f-5459-4178-a8b1-ad249451a47b" containerID="6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846" exitCode=0 Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.354641 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerDied","Data":"6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846"} Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.354667 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerStarted","Data":"56677bdaadc65207c4ea1d260717cc681f1996e18523e70b12cf4e10a3584614"} Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.648895 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4gpsl"] Nov 25 10:23:41 crc kubenswrapper[4926]: W1125 10:23:41.657218 4926 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab59112_b641_43af_a91b_e7c0e58ff417.slice/crio-ef2f38e2df0886a7dcb2be6c43a156a2c70919157df0f6cc4490bcc1d115c0fb WatchSource:0}: Error finding container ef2f38e2df0886a7dcb2be6c43a156a2c70919157df0f6cc4490bcc1d115c0fb: Status 404 returned error can't find the container with id ef2f38e2df0886a7dcb2be6c43a156a2c70919157df0f6cc4490bcc1d115c0fb Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.977772 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rmp4c"] Nov 25 10:23:41 crc kubenswrapper[4926]: I1125 10:23:41.980673 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.038199 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rmp4c"] Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.097743 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-catalog-content\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.097809 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-utilities\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.097849 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bxjf\" (UniqueName: \"kubernetes.io/projected/c8e47d70-d73c-43fe-9a70-7a82c73642c0-kube-api-access-5bxjf\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.199851 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-catalog-content\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.200188 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-utilities\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.200339 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bxjf\" (UniqueName: \"kubernetes.io/projected/c8e47d70-d73c-43fe-9a70-7a82c73642c0-kube-api-access-5bxjf\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.200457 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-catalog-content\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.200904 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-utilities\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.240293 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bxjf\" (UniqueName: \"kubernetes.io/projected/c8e47d70-d73c-43fe-9a70-7a82c73642c0-kube-api-access-5bxjf\") pod \"community-operators-rmp4c\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.323322 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.363585 4926 generic.go:334] "Generic (PLEG): container finished" podID="eab59112-b641-43af-a91b-e7c0e58ff417" containerID="7154c611382956930edaeaa5cf5dffa6a46e67eafc1a01b8b9729fe32f5c53e3" exitCode=0 Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.363649 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerDied","Data":"7154c611382956930edaeaa5cf5dffa6a46e67eafc1a01b8b9729fe32f5c53e3"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.363675 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerStarted","Data":"ef2f38e2df0886a7dcb2be6c43a156a2c70919157df0f6cc4490bcc1d115c0fb"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.365979 4926 generic.go:334] "Generic (PLEG): container finished" podID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerID="11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2" exitCode=0 Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.366031 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerDied","Data":"11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.369418 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerStarted","Data":"82ce0166146e1bbf8ebbc3da9fb34d02030a47ab533510a8c51945287c47c173"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.374259 4926 generic.go:334] "Generic (PLEG): container finished" podID="705621a2-d616-487e-b80a-187d8d6214c2" containerID="6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c" exitCode=0 Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.374355 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jscf7" event={"ID":"705621a2-d616-487e-b80a-187d8d6214c2","Type":"ContainerDied","Data":"6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.376616 4926 generic.go:334] "Generic (PLEG): container finished" podID="a2275402-9d74-4709-9908-add9b0b816c4" containerID="3d8dca83fc7d890bd9d924cdcb3a810cf6b4562d194a7e3135b5e152db20f3e5" exitCode=0 Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.376663 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-527cz" event={"ID":"a2275402-9d74-4709-9908-add9b0b816c4","Type":"ContainerDied","Data":"3d8dca83fc7d890bd9d924cdcb3a810cf6b4562d194a7e3135b5e152db20f3e5"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.379288 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerStarted","Data":"860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.381720 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerDied","Data":"42a8de03ebbfdbb0aeaaa9cf0d2568fc81055cae01217b924427459ee714959e"} Nov 25 10:23:42 crc kubenswrapper[4926]: I1125 10:23:42.381559 4926 generic.go:334] "Generic (PLEG): container finished" podID="55cf4e17-622b-4238-894c-7b3362839762" containerID="42a8de03ebbfdbb0aeaaa9cf0d2568fc81055cae01217b924427459ee714959e" exitCode=0 Nov 25 10:23:43 crc kubenswrapper[4926]: I1125 10:23:43.288215 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rmp4c"] Nov 25 10:23:43 crc kubenswrapper[4926]: I1125 10:23:43.393977 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerStarted","Data":"44b21e4d32314144afd13aa0f0cd47ea64fcd0186a27626c4cc10a0c5d713022"} Nov 25 10:23:44 crc kubenswrapper[4926]: I1125 10:23:44.403401 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerStarted","Data":"464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359"} Nov 25 10:23:44 crc kubenswrapper[4926]: I1125 10:23:44.408370 4926 generic.go:334] "Generic (PLEG): container finished" podID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerID="82ce0166146e1bbf8ebbc3da9fb34d02030a47ab533510a8c51945287c47c173" exitCode=0 Nov 25 10:23:44 crc kubenswrapper[4926]: I1125 10:23:44.408413 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerDied","Data":"82ce0166146e1bbf8ebbc3da9fb34d02030a47ab533510a8c51945287c47c173"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.421316 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerStarted","Data":"fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.429059 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerStarted","Data":"82e502782b5b06fa756ec69f619ce9dffb9ceba17c984c3975e9cd13bff76737"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.439562 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerStarted","Data":"ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.440312 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tcmdr" podStartSLOduration=11.880864953 podStartE2EDuration="15.44030175s" podCreationTimestamp="2025-11-25 10:23:30 +0000 UTC" firstStartedPulling="2025-11-25 10:23:40.292812842 +0000 UTC m=+2116.976769269" lastFinishedPulling="2025-11-25 10:23:43.852249639 +0000 UTC m=+2120.536206066" observedRunningTime="2025-11-25 10:23:45.439300645 +0000 UTC m=+2122.123257082" watchObservedRunningTime="2025-11-25 10:23:45.44030175 +0000 UTC m=+2122.124258167" Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.445865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jscf7" event={"ID":"705621a2-d616-487e-b80a-187d8d6214c2","Type":"ContainerStarted","Data":"8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.447717 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-527cz" event={"ID":"a2275402-9d74-4709-9908-add9b0b816c4","Type":"ContainerStarted","Data":"2d300439e731dda93cc556fd151d229903059f7bda21c9f95b076cd982db9911"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.448851 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerStarted","Data":"0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.456356 4926 generic.go:334] "Generic (PLEG): container finished" podID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerID="860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743" exitCode=0 Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.456923 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerDied","Data":"860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743"} Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.483840 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-527cz" podStartSLOduration=6.753543859 podStartE2EDuration="10.483818671s" podCreationTimestamp="2025-11-25 10:23:35 +0000 UTC" firstStartedPulling="2025-11-25 10:23:40.272388834 +0000 UTC m=+2116.956345271" lastFinishedPulling="2025-11-25 10:23:44.002663656 +0000 UTC m=+2120.686620083" observedRunningTime="2025-11-25 10:23:45.479232047 +0000 UTC m=+2122.163188484" watchObservedRunningTime="2025-11-25 10:23:45.483818671 +0000 UTC m=+2122.167775098" Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.501214 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jscf7" podStartSLOduration=7.930227939 podStartE2EDuration="11.501199123s" podCreationTimestamp="2025-11-25 10:23:34 +0000 UTC" firstStartedPulling="2025-11-25 10:23:40.258257673 +0000 UTC m=+2116.942214100" lastFinishedPulling="2025-11-25 10:23:43.829228857 +0000 UTC m=+2120.513185284" observedRunningTime="2025-11-25 10:23:45.500306521 +0000 UTC m=+2122.184262948" watchObservedRunningTime="2025-11-25 10:23:45.501199123 +0000 UTC m=+2122.185155550" Nov 25 10:23:45 crc kubenswrapper[4926]: I1125 10:23:45.561858 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6phfs" podStartSLOduration=9.117986496 podStartE2EDuration="12.56183311s" podCreationTimestamp="2025-11-25 10:23:33 +0000 UTC" firstStartedPulling="2025-11-25 10:23:40.29876736 +0000 UTC m=+2116.982723787" lastFinishedPulling="2025-11-25 10:23:43.742613974 +0000 UTC m=+2120.426570401" observedRunningTime="2025-11-25 10:23:45.559251426 +0000 UTC m=+2122.243207853" watchObservedRunningTime="2025-11-25 10:23:45.56183311 +0000 UTC m=+2122.245789537" Nov 25 10:23:46 crc kubenswrapper[4926]: I1125 10:23:46.309579 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:46 crc kubenswrapper[4926]: I1125 10:23:46.310002 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:23:46 crc kubenswrapper[4926]: I1125 10:23:46.474133 4926 generic.go:334] "Generic (PLEG): container finished" podID="675b802f-5459-4178-a8b1-ad249451a47b" containerID="464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359" exitCode=0 Nov 25 10:23:46 crc kubenswrapper[4926]: I1125 10:23:46.474211 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerDied","Data":"464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359"} Nov 25 10:23:46 crc kubenswrapper[4926]: I1125 10:23:46.478784 4926 generic.go:334] "Generic (PLEG): container finished" podID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerID="0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2" exitCode=0 Nov 25 10:23:46 crc kubenswrapper[4926]: I1125 10:23:46.478927 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerDied","Data":"0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2"} Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.372051 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-527cz" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="registry-server" probeResult="failure" output=< Nov 25 10:23:47 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:23:47 crc kubenswrapper[4926]: > Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.496277 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerStarted","Data":"d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0"} Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.498971 4926 generic.go:334] "Generic (PLEG): container finished" podID="eab59112-b641-43af-a91b-e7c0e58ff417" containerID="82e502782b5b06fa756ec69f619ce9dffb9ceba17c984c3975e9cd13bff76737" exitCode=0 Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.498984 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerDied","Data":"82e502782b5b06fa756ec69f619ce9dffb9ceba17c984c3975e9cd13bff76737"} Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.508545 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerStarted","Data":"e255ae93dca2d93eddb939e9d2a81e5539352cbb99fa91217b4743758b9c8585"} Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.510362 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.510388 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.535424 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4wgsp" podStartSLOduration=3.609383106 podStartE2EDuration="9.53540146s" podCreationTimestamp="2025-11-25 10:23:38 +0000 UTC" firstStartedPulling="2025-11-25 10:23:40.282725711 +0000 UTC m=+2116.966682138" lastFinishedPulling="2025-11-25 10:23:46.208744065 +0000 UTC m=+2122.892700492" observedRunningTime="2025-11-25 10:23:47.52817932 +0000 UTC m=+2124.212135757" watchObservedRunningTime="2025-11-25 10:23:47.53540146 +0000 UTC m=+2124.219357887" Nov 25 10:23:47 crc kubenswrapper[4926]: I1125 10:23:47.570770 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jdm9g" podStartSLOduration=4.837351481 podStartE2EDuration="10.570751208s" podCreationTimestamp="2025-11-25 10:23:37 +0000 UTC" firstStartedPulling="2025-11-25 10:23:40.306093852 +0000 UTC m=+2116.990050279" lastFinishedPulling="2025-11-25 10:23:46.039493569 +0000 UTC m=+2122.723450006" observedRunningTime="2025-11-25 10:23:47.570324307 +0000 UTC m=+2124.254280734" watchObservedRunningTime="2025-11-25 10:23:47.570751208 +0000 UTC m=+2124.254707635" Nov 25 10:23:48 crc kubenswrapper[4926]: I1125 10:23:48.566215 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jdm9g" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="registry-server" probeResult="failure" output=< Nov 25 10:23:48 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:23:48 crc kubenswrapper[4926]: > Nov 25 10:23:48 crc kubenswrapper[4926]: I1125 10:23:48.763558 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:48 crc kubenswrapper[4926]: I1125 10:23:48.763638 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:49 crc kubenswrapper[4926]: I1125 10:23:49.824631 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4wgsp" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="registry-server" probeResult="failure" output=< Nov 25 10:23:49 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:23:49 crc kubenswrapper[4926]: > Nov 25 10:23:50 crc kubenswrapper[4926]: I1125 10:23:50.695119 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:50 crc kubenswrapper[4926]: I1125 10:23:50.695175 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.553819 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerStarted","Data":"8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4"} Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.556637 4926 generic.go:334] "Generic (PLEG): container finished" podID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerID="9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5" exitCode=0 Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.556705 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerDied","Data":"9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5"} Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.561375 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerStarted","Data":"f5e27cae4bf945cf9a3f376992b790c7cf238ca7dc0362fcfd9e5f3cc0238bd8"} Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.583501 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-stqkg" podStartSLOduration=5.539401714 podStartE2EDuration="12.583481299s" podCreationTimestamp="2025-11-25 10:23:39 +0000 UTC" firstStartedPulling="2025-11-25 10:23:41.378412258 +0000 UTC m=+2118.062368695" lastFinishedPulling="2025-11-25 10:23:48.422491843 +0000 UTC m=+2125.106448280" observedRunningTime="2025-11-25 10:23:51.575132781 +0000 UTC m=+2128.259089228" watchObservedRunningTime="2025-11-25 10:23:51.583481299 +0000 UTC m=+2128.267437726" Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.619735 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4gpsl" podStartSLOduration=5.650290038 podStartE2EDuration="11.619715589s" podCreationTimestamp="2025-11-25 10:23:40 +0000 UTC" firstStartedPulling="2025-11-25 10:23:42.464502576 +0000 UTC m=+2119.148459003" lastFinishedPulling="2025-11-25 10:23:48.433928117 +0000 UTC m=+2125.117884554" observedRunningTime="2025-11-25 10:23:51.613474544 +0000 UTC m=+2128.297430981" watchObservedRunningTime="2025-11-25 10:23:51.619715589 +0000 UTC m=+2128.303672016" Nov 25 10:23:51 crc kubenswrapper[4926]: I1125 10:23:51.753694 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-tcmdr" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="registry-server" probeResult="failure" output=< Nov 25 10:23:51 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:23:51 crc kubenswrapper[4926]: > Nov 25 10:23:52 crc kubenswrapper[4926]: I1125 10:23:52.573326 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerStarted","Data":"80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67"} Nov 25 10:23:52 crc kubenswrapper[4926]: I1125 10:23:52.610640 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rmp4c" podStartSLOduration=5.858873289 podStartE2EDuration="11.610623292s" podCreationTimestamp="2025-11-25 10:23:41 +0000 UTC" firstStartedPulling="2025-11-25 10:23:46.480408054 +0000 UTC m=+2123.164364491" lastFinishedPulling="2025-11-25 10:23:52.232158067 +0000 UTC m=+2128.916114494" observedRunningTime="2025-11-25 10:23:52.602199852 +0000 UTC m=+2129.286156279" watchObservedRunningTime="2025-11-25 10:23:52.610623292 +0000 UTC m=+2129.294579719" Nov 25 10:23:53 crc kubenswrapper[4926]: I1125 10:23:53.732571 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:53 crc kubenswrapper[4926]: I1125 10:23:53.736175 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:23:54 crc kubenswrapper[4926]: I1125 10:23:54.593626 4926 generic.go:334] "Generic (PLEG): container finished" podID="097dab84-8b0b-4bbc-af0f-18b062a1a891" containerID="a030c75c8c794bfe9a30303013f9354277d780fa603207a12ab49243501550a3" exitCode=0 Nov 25 10:23:54 crc kubenswrapper[4926]: I1125 10:23:54.593697 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" event={"ID":"097dab84-8b0b-4bbc-af0f-18b062a1a891","Type":"ContainerDied","Data":"a030c75c8c794bfe9a30303013f9354277d780fa603207a12ab49243501550a3"} Nov 25 10:23:54 crc kubenswrapper[4926]: I1125 10:23:54.813685 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6phfs" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="registry-server" probeResult="failure" output=< Nov 25 10:23:54 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:23:54 crc kubenswrapper[4926]: > Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.106922 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.106996 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.158206 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.681637 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.732055 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.777132 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wtrhn/crc-debug-zw6x8"] Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.790151 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wtrhn/crc-debug-zw6x8"] Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.846279 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlcqf\" (UniqueName: \"kubernetes.io/projected/097dab84-8b0b-4bbc-af0f-18b062a1a891-kube-api-access-jlcqf\") pod \"097dab84-8b0b-4bbc-af0f-18b062a1a891\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.846412 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/097dab84-8b0b-4bbc-af0f-18b062a1a891-host\") pod \"097dab84-8b0b-4bbc-af0f-18b062a1a891\" (UID: \"097dab84-8b0b-4bbc-af0f-18b062a1a891\") " Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.846542 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/097dab84-8b0b-4bbc-af0f-18b062a1a891-host" (OuterVolumeSpecName: "host") pod "097dab84-8b0b-4bbc-af0f-18b062a1a891" (UID: "097dab84-8b0b-4bbc-af0f-18b062a1a891"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.846996 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/097dab84-8b0b-4bbc-af0f-18b062a1a891-host\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.866823 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/097dab84-8b0b-4bbc-af0f-18b062a1a891-kube-api-access-jlcqf" (OuterVolumeSpecName: "kube-api-access-jlcqf") pod "097dab84-8b0b-4bbc-af0f-18b062a1a891" (UID: "097dab84-8b0b-4bbc-af0f-18b062a1a891"). InnerVolumeSpecName "kube-api-access-jlcqf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:23:55 crc kubenswrapper[4926]: I1125 10:23:55.948133 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlcqf\" (UniqueName: \"kubernetes.io/projected/097dab84-8b0b-4bbc-af0f-18b062a1a891-kube-api-access-jlcqf\") on node \"crc\" DevicePath \"\"" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.023756 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="097dab84-8b0b-4bbc-af0f-18b062a1a891" path="/var/lib/kubelet/pods/097dab84-8b0b-4bbc-af0f-18b062a1a891/volumes" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.618967 4926 scope.go:117] "RemoveContainer" containerID="a030c75c8c794bfe9a30303013f9354277d780fa603207a12ab49243501550a3" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.618991 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-zw6x8" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.992177 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-wtrhn/crc-debug-sktm5"] Nov 25 10:23:56 crc kubenswrapper[4926]: E1125 10:23:56.992548 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="097dab84-8b0b-4bbc-af0f-18b062a1a891" containerName="container-00" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.992560 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="097dab84-8b0b-4bbc-af0f-18b062a1a891" containerName="container-00" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.992748 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="097dab84-8b0b-4bbc-af0f-18b062a1a891" containerName="container-00" Nov 25 10:23:56 crc kubenswrapper[4926]: I1125 10:23:56.993328 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.178356 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-host\") pod \"crc-debug-sktm5\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.178898 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbvcf\" (UniqueName: \"kubernetes.io/projected/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-kube-api-access-pbvcf\") pod \"crc-debug-sktm5\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.280645 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-host\") pod \"crc-debug-sktm5\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.280802 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbvcf\" (UniqueName: \"kubernetes.io/projected/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-kube-api-access-pbvcf\") pod \"crc-debug-sktm5\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.280802 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-host\") pod \"crc-debug-sktm5\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.319552 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbvcf\" (UniqueName: \"kubernetes.io/projected/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-kube-api-access-pbvcf\") pod \"crc-debug-sktm5\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.376428 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-527cz" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="registry-server" probeResult="failure" output=< Nov 25 10:23:57 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:23:57 crc kubenswrapper[4926]: > Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.568816 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.613764 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:23:57 crc kubenswrapper[4926]: I1125 10:23:57.635296 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.640155 4926 generic.go:334] "Generic (PLEG): container finished" podID="ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" containerID="9ea075134112737ea4398b1fd34e52d0e48fcec09fee71009604c69d9813f247" exitCode=1 Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.640282 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/crc-debug-sktm5" event={"ID":"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0","Type":"ContainerDied","Data":"9ea075134112737ea4398b1fd34e52d0e48fcec09fee71009604c69d9813f247"} Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.640889 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/crc-debug-sktm5" event={"ID":"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0","Type":"ContainerStarted","Data":"542417083e88ad0a5ac818a1c27a9364deeaf6dde03608554d4dda5ba23a9398"} Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.691257 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wtrhn/crc-debug-sktm5"] Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.700786 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wtrhn/crc-debug-sktm5"] Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.816249 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:23:58 crc kubenswrapper[4926]: I1125 10:23:58.862346 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.582439 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.582885 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.676858 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="542417083e88ad0a5ac818a1c27a9364deeaf6dde03608554d4dda5ba23a9398" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.756552 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.802892 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.828347 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.882206 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-host\") pod \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.882322 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pbvcf\" (UniqueName: \"kubernetes.io/projected/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-kube-api-access-pbvcf\") pod \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\" (UID: \"ae4adb9a-0c49-4c55-ba5c-1483c563c1d0\") " Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.882584 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-host" (OuterVolumeSpecName: "host") pod "ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" (UID: "ae4adb9a-0c49-4c55-ba5c-1483c563c1d0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.888296 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-kube-api-access-pbvcf" (OuterVolumeSpecName: "kube-api-access-pbvcf") pod "ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" (UID: "ae4adb9a-0c49-4c55-ba5c-1483c563c1d0"). InnerVolumeSpecName "kube-api-access-pbvcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.984148 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pbvcf\" (UniqueName: \"kubernetes.io/projected/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-kube-api-access-pbvcf\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:00 crc kubenswrapper[4926]: I1125 10:24:00.984186 4926 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0-host\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:01 crc kubenswrapper[4926]: I1125 10:24:01.138335 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:24:01 crc kubenswrapper[4926]: I1125 10:24:01.138398 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:24:01 crc kubenswrapper[4926]: I1125 10:24:01.684300 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/crc-debug-sktm5" Nov 25 10:24:01 crc kubenswrapper[4926]: I1125 10:24:01.724370 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-stqkg" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="registry-server" probeResult="failure" output=< Nov 25 10:24:01 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:24:01 crc kubenswrapper[4926]: > Nov 25 10:24:02 crc kubenswrapper[4926]: I1125 10:24:02.032540 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" path="/var/lib/kubelet/pods/ae4adb9a-0c49-4c55-ba5c-1483c563c1d0/volumes" Nov 25 10:24:02 crc kubenswrapper[4926]: I1125 10:24:02.182565 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4gpsl" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="registry-server" probeResult="failure" output=< Nov 25 10:24:02 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:24:02 crc kubenswrapper[4926]: > Nov 25 10:24:02 crc kubenswrapper[4926]: I1125 10:24:02.324411 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:24:02 crc kubenswrapper[4926]: I1125 10:24:02.324462 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:24:03 crc kubenswrapper[4926]: I1125 10:24:03.375485 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rmp4c" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="registry-server" probeResult="failure" output=< Nov 25 10:24:03 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:24:03 crc kubenswrapper[4926]: > Nov 25 10:24:03 crc kubenswrapper[4926]: I1125 10:24:03.780818 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:24:03 crc kubenswrapper[4926]: I1125 10:24:03.839843 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:24:06 crc kubenswrapper[4926]: I1125 10:24:06.359277 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:24:06 crc kubenswrapper[4926]: I1125 10:24:06.411663 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:24:11 crc kubenswrapper[4926]: I1125 10:24:11.009568 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-stqkg" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="registry-server" probeResult="failure" output=< Nov 25 10:24:11 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:24:11 crc kubenswrapper[4926]: > Nov 25 10:24:12 crc kubenswrapper[4926]: I1125 10:24:12.213447 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4gpsl" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="registry-server" probeResult="failure" output=< Nov 25 10:24:12 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:24:12 crc kubenswrapper[4926]: > Nov 25 10:24:13 crc kubenswrapper[4926]: I1125 10:24:13.367707 4926 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-rmp4c" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="registry-server" probeResult="failure" output=< Nov 25 10:24:13 crc kubenswrapper[4926]: timeout: failed to connect service ":50051" within 1s Nov 25 10:24:13 crc kubenswrapper[4926]: > Nov 25 10:24:14 crc kubenswrapper[4926]: I1125 10:24:14.023391 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:24:14 crc kubenswrapper[4926]: I1125 10:24:14.023447 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:24:20 crc kubenswrapper[4926]: I1125 10:24:20.022250 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:24:20 crc kubenswrapper[4926]: I1125 10:24:20.077615 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:24:21 crc kubenswrapper[4926]: I1125 10:24:21.191833 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:24:21 crc kubenswrapper[4926]: I1125 10:24:21.246947 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:24:22 crc kubenswrapper[4926]: I1125 10:24:22.366799 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:24:22 crc kubenswrapper[4926]: I1125 10:24:22.422534 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:24:23 crc kubenswrapper[4926]: I1125 10:24:23.770390 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4gpsl"] Nov 25 10:24:23 crc kubenswrapper[4926]: I1125 10:24:23.770891 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4gpsl" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="registry-server" containerID="cri-o://f5e27cae4bf945cf9a3f376992b790c7cf238ca7dc0362fcfd9e5f3cc0238bd8" gracePeriod=2 Nov 25 10:24:23 crc kubenswrapper[4926]: I1125 10:24:23.899751 4926 generic.go:334] "Generic (PLEG): container finished" podID="eab59112-b641-43af-a91b-e7c0e58ff417" containerID="f5e27cae4bf945cf9a3f376992b790c7cf238ca7dc0362fcfd9e5f3cc0238bd8" exitCode=0 Nov 25 10:24:23 crc kubenswrapper[4926]: I1125 10:24:23.899801 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerDied","Data":"f5e27cae4bf945cf9a3f376992b790c7cf238ca7dc0362fcfd9e5f3cc0238bd8"} Nov 25 10:24:23 crc kubenswrapper[4926]: I1125 10:24:23.970948 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4wgsp"] Nov 25 10:24:23 crc kubenswrapper[4926]: I1125 10:24:23.971250 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4wgsp" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="registry-server" containerID="cri-o://d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0" gracePeriod=2 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.247636 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rmp4c"] Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.248327 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rmp4c" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="registry-server" containerID="cri-o://80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67" gracePeriod=2 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.486471 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.512169 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.571905 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-527cz"] Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.572212 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-527cz" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="registry-server" containerID="cri-o://2d300439e731dda93cc556fd151d229903059f7bda21c9f95b076cd982db9911" gracePeriod=2 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.676847 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-utilities\") pod \"25b0e9bd-d961-4465-9496-14420a9daf2c\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.677007 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-utilities\") pod \"eab59112-b641-43af-a91b-e7c0e58ff417\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.677139 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnrjj\" (UniqueName: \"kubernetes.io/projected/eab59112-b641-43af-a91b-e7c0e58ff417-kube-api-access-dnrjj\") pod \"eab59112-b641-43af-a91b-e7c0e58ff417\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.677176 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-catalog-content\") pod \"25b0e9bd-d961-4465-9496-14420a9daf2c\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.677236 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7294z\" (UniqueName: \"kubernetes.io/projected/25b0e9bd-d961-4465-9496-14420a9daf2c-kube-api-access-7294z\") pod \"25b0e9bd-d961-4465-9496-14420a9daf2c\" (UID: \"25b0e9bd-d961-4465-9496-14420a9daf2c\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.677286 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-catalog-content\") pod \"eab59112-b641-43af-a91b-e7c0e58ff417\" (UID: \"eab59112-b641-43af-a91b-e7c0e58ff417\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.677788 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-utilities" (OuterVolumeSpecName: "utilities") pod "25b0e9bd-d961-4465-9496-14420a9daf2c" (UID: "25b0e9bd-d961-4465-9496-14420a9daf2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.682388 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25b0e9bd-d961-4465-9496-14420a9daf2c-kube-api-access-7294z" (OuterVolumeSpecName: "kube-api-access-7294z") pod "25b0e9bd-d961-4465-9496-14420a9daf2c" (UID: "25b0e9bd-d961-4465-9496-14420a9daf2c"). InnerVolumeSpecName "kube-api-access-7294z". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.686708 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-utilities" (OuterVolumeSpecName: "utilities") pod "eab59112-b641-43af-a91b-e7c0e58ff417" (UID: "eab59112-b641-43af-a91b-e7c0e58ff417"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.690678 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eab59112-b641-43af-a91b-e7c0e58ff417-kube-api-access-dnrjj" (OuterVolumeSpecName: "kube-api-access-dnrjj") pod "eab59112-b641-43af-a91b-e7c0e58ff417" (UID: "eab59112-b641-43af-a91b-e7c0e58ff417"). InnerVolumeSpecName "kube-api-access-dnrjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.722687 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.750708 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25b0e9bd-d961-4465-9496-14420a9daf2c" (UID: "25b0e9bd-d961-4465-9496-14420a9daf2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.754659 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eab59112-b641-43af-a91b-e7c0e58ff417" (UID: "eab59112-b641-43af-a91b-e7c0e58ff417"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.778851 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-catalog-content\") pod \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.778984 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bxjf\" (UniqueName: \"kubernetes.io/projected/c8e47d70-d73c-43fe-9a70-7a82c73642c0-kube-api-access-5bxjf\") pod \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779014 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-utilities\") pod \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\" (UID: \"c8e47d70-d73c-43fe-9a70-7a82c73642c0\") " Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779284 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779295 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnrjj\" (UniqueName: \"kubernetes.io/projected/eab59112-b641-43af-a91b-e7c0e58ff417-kube-api-access-dnrjj\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779305 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779314 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7294z\" (UniqueName: \"kubernetes.io/projected/25b0e9bd-d961-4465-9496-14420a9daf2c-kube-api-access-7294z\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779322 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eab59112-b641-43af-a91b-e7c0e58ff417-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.779330 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25b0e9bd-d961-4465-9496-14420a9daf2c-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.780071 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-utilities" (OuterVolumeSpecName: "utilities") pod "c8e47d70-d73c-43fe-9a70-7a82c73642c0" (UID: "c8e47d70-d73c-43fe-9a70-7a82c73642c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.783267 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8e47d70-d73c-43fe-9a70-7a82c73642c0-kube-api-access-5bxjf" (OuterVolumeSpecName: "kube-api-access-5bxjf") pod "c8e47d70-d73c-43fe-9a70-7a82c73642c0" (UID: "c8e47d70-d73c-43fe-9a70-7a82c73642c0"). InnerVolumeSpecName "kube-api-access-5bxjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.838992 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c8e47d70-d73c-43fe-9a70-7a82c73642c0" (UID: "c8e47d70-d73c-43fe-9a70-7a82c73642c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.884528 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.884579 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bxjf\" (UniqueName: \"kubernetes.io/projected/c8e47d70-d73c-43fe-9a70-7a82c73642c0-kube-api-access-5bxjf\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.884600 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c8e47d70-d73c-43fe-9a70-7a82c73642c0-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.925350 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4gpsl" event={"ID":"eab59112-b641-43af-a91b-e7c0e58ff417","Type":"ContainerDied","Data":"ef2f38e2df0886a7dcb2be6c43a156a2c70919157df0f6cc4490bcc1d115c0fb"} Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.925487 4926 scope.go:117] "RemoveContainer" containerID="f5e27cae4bf945cf9a3f376992b790c7cf238ca7dc0362fcfd9e5f3cc0238bd8" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.925641 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4gpsl" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.935538 4926 generic.go:334] "Generic (PLEG): container finished" podID="a2275402-9d74-4709-9908-add9b0b816c4" containerID="2d300439e731dda93cc556fd151d229903059f7bda21c9f95b076cd982db9911" exitCode=0 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.935610 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-527cz" event={"ID":"a2275402-9d74-4709-9908-add9b0b816c4","Type":"ContainerDied","Data":"2d300439e731dda93cc556fd151d229903059f7bda21c9f95b076cd982db9911"} Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.938061 4926 generic.go:334] "Generic (PLEG): container finished" podID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerID="80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67" exitCode=0 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.938122 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerDied","Data":"80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67"} Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.938135 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rmp4c" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.938151 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rmp4c" event={"ID":"c8e47d70-d73c-43fe-9a70-7a82c73642c0","Type":"ContainerDied","Data":"44b21e4d32314144afd13aa0f0cd47ea64fcd0186a27626c4cc10a0c5d713022"} Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.947323 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4wgsp" Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.947258 4926 generic.go:334] "Generic (PLEG): container finished" podID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerID="d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0" exitCode=0 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.947360 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerDied","Data":"d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0"} Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.947979 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4wgsp" event={"ID":"25b0e9bd-d961-4465-9496-14420a9daf2c","Type":"ContainerDied","Data":"b1cd4a6da2d7b2f769e58f9825bbec5b222398045b7e2550804a3c001fb013a7"} Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.980912 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-stqkg"] Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.981124 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-stqkg" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="registry-server" containerID="cri-o://8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4" gracePeriod=2 Nov 25 10:24:24 crc kubenswrapper[4926]: I1125 10:24:24.998622 4926 scope.go:117] "RemoveContainer" containerID="82e502782b5b06fa756ec69f619ce9dffb9ceba17c984c3975e9cd13bff76737" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.003981 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4gpsl"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.014738 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4gpsl"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.223742 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4wgsp"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.230592 4926 scope.go:117] "RemoveContainer" containerID="7154c611382956930edaeaa5cf5dffa6a46e67eafc1a01b8b9729fe32f5c53e3" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.236810 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4wgsp"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.249660 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.257084 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rmp4c"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.267823 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rmp4c"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.284553 4926 scope.go:117] "RemoveContainer" containerID="80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.326300 4926 scope.go:117] "RemoveContainer" containerID="9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.354247 4926 scope.go:117] "RemoveContainer" containerID="0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.371695 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.376378 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6phfs"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.376631 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6phfs" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="registry-server" containerID="cri-o://ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2" gracePeriod=2 Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.379203 4926 scope.go:117] "RemoveContainer" containerID="80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67" Nov 25 10:24:25 crc kubenswrapper[4926]: E1125 10:24:25.379762 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67\": container with ID starting with 80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67 not found: ID does not exist" containerID="80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.379788 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67"} err="failed to get container status \"80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67\": rpc error: code = NotFound desc = could not find container \"80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67\": container with ID starting with 80a574aba5bd401e073d6a37b3986c7637e57a25114268f5fa18ccbe8ffa9b67 not found: ID does not exist" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.379806 4926 scope.go:117] "RemoveContainer" containerID="9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5" Nov 25 10:24:25 crc kubenswrapper[4926]: E1125 10:24:25.380240 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5\": container with ID starting with 9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5 not found: ID does not exist" containerID="9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.380299 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5"} err="failed to get container status \"9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5\": rpc error: code = NotFound desc = could not find container \"9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5\": container with ID starting with 9939384b404107ee8b01bbb79458b06835dd9715a4c93ba4c2b36c493444cfa5 not found: ID does not exist" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.380322 4926 scope.go:117] "RemoveContainer" containerID="0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2" Nov 25 10:24:25 crc kubenswrapper[4926]: E1125 10:24:25.381897 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2\": container with ID starting with 0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2 not found: ID does not exist" containerID="0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.381946 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2"} err="failed to get container status \"0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2\": rpc error: code = NotFound desc = could not find container \"0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2\": container with ID starting with 0cc740fc27cf95cb6684816c4bdf804fa879888a188e3b53d93739cc6d9c57d2 not found: ID does not exist" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.381969 4926 scope.go:117] "RemoveContainer" containerID="d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.399160 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q7gv\" (UniqueName: \"kubernetes.io/projected/675b802f-5459-4178-a8b1-ad249451a47b-kube-api-access-5q7gv\") pod \"675b802f-5459-4178-a8b1-ad249451a47b\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.399233 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sjv2\" (UniqueName: \"kubernetes.io/projected/a2275402-9d74-4709-9908-add9b0b816c4-kube-api-access-9sjv2\") pod \"a2275402-9d74-4709-9908-add9b0b816c4\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.399256 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-catalog-content\") pod \"a2275402-9d74-4709-9908-add9b0b816c4\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.399292 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-utilities\") pod \"675b802f-5459-4178-a8b1-ad249451a47b\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.399331 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-catalog-content\") pod \"675b802f-5459-4178-a8b1-ad249451a47b\" (UID: \"675b802f-5459-4178-a8b1-ad249451a47b\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.399357 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-utilities\") pod \"a2275402-9d74-4709-9908-add9b0b816c4\" (UID: \"a2275402-9d74-4709-9908-add9b0b816c4\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.400217 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-utilities" (OuterVolumeSpecName: "utilities") pod "a2275402-9d74-4709-9908-add9b0b816c4" (UID: "a2275402-9d74-4709-9908-add9b0b816c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.404439 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-utilities" (OuterVolumeSpecName: "utilities") pod "675b802f-5459-4178-a8b1-ad249451a47b" (UID: "675b802f-5459-4178-a8b1-ad249451a47b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.406232 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2275402-9d74-4709-9908-add9b0b816c4-kube-api-access-9sjv2" (OuterVolumeSpecName: "kube-api-access-9sjv2") pod "a2275402-9d74-4709-9908-add9b0b816c4" (UID: "a2275402-9d74-4709-9908-add9b0b816c4"). InnerVolumeSpecName "kube-api-access-9sjv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.411099 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675b802f-5459-4178-a8b1-ad249451a47b-kube-api-access-5q7gv" (OuterVolumeSpecName: "kube-api-access-5q7gv") pod "675b802f-5459-4178-a8b1-ad249451a47b" (UID: "675b802f-5459-4178-a8b1-ad249451a47b"). InnerVolumeSpecName "kube-api-access-5q7gv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.412898 4926 scope.go:117] "RemoveContainer" containerID="860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.448096 4926 scope.go:117] "RemoveContainer" containerID="4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.468628 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "675b802f-5459-4178-a8b1-ad249451a47b" (UID: "675b802f-5459-4178-a8b1-ad249451a47b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.472727 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a2275402-9d74-4709-9908-add9b0b816c4" (UID: "a2275402-9d74-4709-9908-add9b0b816c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.500777 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.500812 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q7gv\" (UniqueName: \"kubernetes.io/projected/675b802f-5459-4178-a8b1-ad249451a47b-kube-api-access-5q7gv\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.500823 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sjv2\" (UniqueName: \"kubernetes.io/projected/a2275402-9d74-4709-9908-add9b0b816c4-kube-api-access-9sjv2\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.501090 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a2275402-9d74-4709-9908-add9b0b816c4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.501110 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.501121 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/675b802f-5459-4178-a8b1-ad249451a47b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.521268 4926 scope.go:117] "RemoveContainer" containerID="d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0" Nov 25 10:24:25 crc kubenswrapper[4926]: E1125 10:24:25.522018 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0\": container with ID starting with d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0 not found: ID does not exist" containerID="d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.522064 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0"} err="failed to get container status \"d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0\": rpc error: code = NotFound desc = could not find container \"d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0\": container with ID starting with d5181f3ddab85746eb319b5b50995db6bb17226c681c687cf057f84b53605fe0 not found: ID does not exist" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.522094 4926 scope.go:117] "RemoveContainer" containerID="860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743" Nov 25 10:24:25 crc kubenswrapper[4926]: E1125 10:24:25.522901 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743\": container with ID starting with 860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743 not found: ID does not exist" containerID="860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.522955 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743"} err="failed to get container status \"860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743\": rpc error: code = NotFound desc = could not find container \"860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743\": container with ID starting with 860b19bdbd6b1c75f4647000b52715edf8f2bd1c5a9bb1b748ad719197567743 not found: ID does not exist" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.522982 4926 scope.go:117] "RemoveContainer" containerID="4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53" Nov 25 10:24:25 crc kubenswrapper[4926]: E1125 10:24:25.523434 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53\": container with ID starting with 4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53 not found: ID does not exist" containerID="4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.523455 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53"} err="failed to get container status \"4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53\": rpc error: code = NotFound desc = could not find container \"4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53\": container with ID starting with 4c88a06ec31ea92c367dbb9f9cf8890787ea31c1a4ff8b9de6117dac715f4f53 not found: ID does not exist" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.745003 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.784466 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jdm9g"] Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.785139 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jdm9g" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="registry-server" containerID="cri-o://e255ae93dca2d93eddb939e9d2a81e5539352cbb99fa91217b4743758b9c8585" gracePeriod=2 Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.906476 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgf7w\" (UniqueName: \"kubernetes.io/projected/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-kube-api-access-rgf7w\") pod \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.906532 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-utilities\") pod \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.906724 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-catalog-content\") pod \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\" (UID: \"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf\") " Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.908119 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-utilities" (OuterVolumeSpecName: "utilities") pod "1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" (UID: "1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.908512 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.912548 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-kube-api-access-rgf7w" (OuterVolumeSpecName: "kube-api-access-rgf7w") pod "1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" (UID: "1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf"). InnerVolumeSpecName "kube-api-access-rgf7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.961034 4926 generic.go:334] "Generic (PLEG): container finished" podID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerID="ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2" exitCode=0 Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.961104 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerDied","Data":"ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2"} Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.961137 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6phfs" event={"ID":"1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf","Type":"ContainerDied","Data":"3e65fac444114c7cf63e5eb1176a63807ace82c6083fb60d6e2d909e6d56409c"} Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.961156 4926 scope.go:117] "RemoveContainer" containerID="ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.961323 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6phfs" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.966083 4926 generic.go:334] "Generic (PLEG): container finished" podID="675b802f-5459-4178-a8b1-ad249451a47b" containerID="8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4" exitCode=0 Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.966254 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-stqkg" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.971459 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" (UID: "1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.971527 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerDied","Data":"8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4"} Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.971562 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-stqkg" event={"ID":"675b802f-5459-4178-a8b1-ad249451a47b","Type":"ContainerDied","Data":"56677bdaadc65207c4ea1d260717cc681f1996e18523e70b12cf4e10a3584614"} Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.975759 4926 generic.go:334] "Generic (PLEG): container finished" podID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerID="e255ae93dca2d93eddb939e9d2a81e5539352cbb99fa91217b4743758b9c8585" exitCode=0 Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.975806 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerDied","Data":"e255ae93dca2d93eddb939e9d2a81e5539352cbb99fa91217b4743758b9c8585"} Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.977571 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-527cz" event={"ID":"a2275402-9d74-4709-9908-add9b0b816c4","Type":"ContainerDied","Data":"cf9b8bbd48302c671c01616fe3c04880ac9cea82a51014a974a0636a2df79402"} Nov 25 10:24:25 crc kubenswrapper[4926]: I1125 10:24:25.977655 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-527cz" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.009573 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgf7w\" (UniqueName: \"kubernetes.io/projected/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-kube-api-access-rgf7w\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.009600 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.042892 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" path="/var/lib/kubelet/pods/25b0e9bd-d961-4465-9496-14420a9daf2c/volumes" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.043632 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" path="/var/lib/kubelet/pods/c8e47d70-d73c-43fe-9a70-7a82c73642c0/volumes" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.044412 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" path="/var/lib/kubelet/pods/eab59112-b641-43af-a91b-e7c0e58ff417/volumes" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.045840 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-stqkg"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.051895 4926 scope.go:117] "RemoveContainer" containerID="11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.058859 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-stqkg"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.070377 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-527cz"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.076689 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-527cz"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.119161 4926 scope.go:117] "RemoveContainer" containerID="80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.156556 4926 scope.go:117] "RemoveContainer" containerID="ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.160561 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2\": container with ID starting with ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2 not found: ID does not exist" containerID="ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.160624 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2"} err="failed to get container status \"ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2\": rpc error: code = NotFound desc = could not find container \"ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2\": container with ID starting with ef3a0332d975d9e83474e5964b576069e4fd4195ddc3fe62a69aa31724241af2 not found: ID does not exist" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.160652 4926 scope.go:117] "RemoveContainer" containerID="11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.162925 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2\": container with ID starting with 11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2 not found: ID does not exist" containerID="11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.163017 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2"} err="failed to get container status \"11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2\": rpc error: code = NotFound desc = could not find container \"11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2\": container with ID starting with 11c86953d9341f2f807803f1b3b2ce51bdb619d98ac97b6d4ae460c1664312e2 not found: ID does not exist" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.163046 4926 scope.go:117] "RemoveContainer" containerID="80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.166561 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585\": container with ID starting with 80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585 not found: ID does not exist" containerID="80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.166602 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585"} err="failed to get container status \"80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585\": rpc error: code = NotFound desc = could not find container \"80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585\": container with ID starting with 80a752b29e2ebc594c9368d56855acf5b1c18d430e00c0847513bc326a10e585 not found: ID does not exist" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.166628 4926 scope.go:117] "RemoveContainer" containerID="8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.179709 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jscf7"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.179988 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jscf7" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="registry-server" containerID="cri-o://8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9" gracePeriod=2 Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.303214 4926 scope.go:117] "RemoveContainer" containerID="464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.316817 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.330902 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6phfs"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.334798 4926 scope.go:117] "RemoveContainer" containerID="6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.343134 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6phfs"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.361398 4926 scope.go:117] "RemoveContainer" containerID="8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.361905 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4\": container with ID starting with 8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4 not found: ID does not exist" containerID="8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.362011 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4"} err="failed to get container status \"8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4\": rpc error: code = NotFound desc = could not find container \"8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4\": container with ID starting with 8510141bfed7592c4000d2af5e1db670adc50d852d42afd0ad7e8788d081bca4 not found: ID does not exist" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.362036 4926 scope.go:117] "RemoveContainer" containerID="464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.362280 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359\": container with ID starting with 464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359 not found: ID does not exist" containerID="464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.362296 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359"} err="failed to get container status \"464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359\": rpc error: code = NotFound desc = could not find container \"464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359\": container with ID starting with 464de9c499cae7541b6629bb82b1c26d331f98012fd524209b4def70a8912359 not found: ID does not exist" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.362309 4926 scope.go:117] "RemoveContainer" containerID="6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.364134 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846\": container with ID starting with 6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846 not found: ID does not exist" containerID="6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.364160 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846"} err="failed to get container status \"6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846\": rpc error: code = NotFound desc = could not find container \"6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846\": container with ID starting with 6c1a85041424590e9a5fa0905d13ceaaae74eafed9cb29fa5456eaefc7e04846 not found: ID does not exist" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.364175 4926 scope.go:117] "RemoveContainer" containerID="2d300439e731dda93cc556fd151d229903059f7bda21c9f95b076cd982db9911" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.418820 4926 scope.go:117] "RemoveContainer" containerID="3d8dca83fc7d890bd9d924cdcb3a810cf6b4562d194a7e3135b5e152db20f3e5" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.438137 4926 scope.go:117] "RemoveContainer" containerID="e91972f16c3ba7fe0aff40b29213e8cfb5e9306b13a6d2501175fb712636ac84" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.516293 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-86x69\" (UniqueName: \"kubernetes.io/projected/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-kube-api-access-86x69\") pod \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.516345 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-utilities\") pod \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.516454 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-catalog-content\") pod \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\" (UID: \"1934d666-ff68-48e3-8c8e-1dbbcb39cf26\") " Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.517799 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-utilities" (OuterVolumeSpecName: "utilities") pod "1934d666-ff68-48e3-8c8e-1dbbcb39cf26" (UID: "1934d666-ff68-48e3-8c8e-1dbbcb39cf26"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.521748 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-kube-api-access-86x69" (OuterVolumeSpecName: "kube-api-access-86x69") pod "1934d666-ff68-48e3-8c8e-1dbbcb39cf26" (UID: "1934d666-ff68-48e3-8c8e-1dbbcb39cf26"). InnerVolumeSpecName "kube-api-access-86x69". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.574956 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tcmdr"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.575240 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-tcmdr" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="registry-server" containerID="cri-o://fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71" gracePeriod=2 Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.604511 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.616580 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1934d666-ff68-48e3-8c8e-1dbbcb39cf26" (UID: "1934d666-ff68-48e3-8c8e-1dbbcb39cf26"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.619758 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-86x69\" (UniqueName: \"kubernetes.io/projected/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-kube-api-access-86x69\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.619790 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.619802 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1934d666-ff68-48e3-8c8e-1dbbcb39cf26-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.720490 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2s2c\" (UniqueName: \"kubernetes.io/projected/705621a2-d616-487e-b80a-187d8d6214c2-kube-api-access-l2s2c\") pod \"705621a2-d616-487e-b80a-187d8d6214c2\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.720709 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-catalog-content\") pod \"705621a2-d616-487e-b80a-187d8d6214c2\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.720729 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-utilities\") pod \"705621a2-d616-487e-b80a-187d8d6214c2\" (UID: \"705621a2-d616-487e-b80a-187d8d6214c2\") " Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.721774 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-utilities" (OuterVolumeSpecName: "utilities") pod "705621a2-d616-487e-b80a-187d8d6214c2" (UID: "705621a2-d616-487e-b80a-187d8d6214c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.727361 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/705621a2-d616-487e-b80a-187d8d6214c2-kube-api-access-l2s2c" (OuterVolumeSpecName: "kube-api-access-l2s2c") pod "705621a2-d616-487e-b80a-187d8d6214c2" (UID: "705621a2-d616-487e-b80a-187d8d6214c2"). InnerVolumeSpecName "kube-api-access-l2s2c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.779602 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "705621a2-d616-487e-b80a-187d8d6214c2" (UID: "705621a2-d616-487e-b80a-187d8d6214c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:26 crc kubenswrapper[4926]: E1125 10:24:26.805779 4926 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55cf4e17_622b_4238_894c_7b3362839762.slice/crio-fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55cf4e17_622b_4238_894c_7b3362839762.slice/crio-conmon-fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71.scope\": RecentStats: unable to find data in memory cache]" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.823054 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2s2c\" (UniqueName: \"kubernetes.io/projected/705621a2-d616-487e-b80a-187d8d6214c2-kube-api-access-l2s2c\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.823094 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.823103 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/705621a2-d616-487e-b80a-187d8d6214c2-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.978455 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vzjll"] Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.979112 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-vzjll" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="registry-server" containerID="cri-o://bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b" gracePeriod=2 Nov 25 10:24:26 crc kubenswrapper[4926]: I1125 10:24:26.996436 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jdm9g" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:26.999984 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jdm9g" event={"ID":"1934d666-ff68-48e3-8c8e-1dbbcb39cf26","Type":"ContainerDied","Data":"0361a102815644dbdd61a2ab70d92d12bb42d01475df714a3cf494defb923d1f"} Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.000042 4926 scope.go:117] "RemoveContainer" containerID="e255ae93dca2d93eddb939e9d2a81e5539352cbb99fa91217b4743758b9c8585" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.004103 4926 generic.go:334] "Generic (PLEG): container finished" podID="705621a2-d616-487e-b80a-187d8d6214c2" containerID="8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9" exitCode=0 Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.004157 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jscf7" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.004202 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jscf7" event={"ID":"705621a2-d616-487e-b80a-187d8d6214c2","Type":"ContainerDied","Data":"8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9"} Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.004251 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jscf7" event={"ID":"705621a2-d616-487e-b80a-187d8d6214c2","Type":"ContainerDied","Data":"eb4675e24ac0a438940812bce1372c24f9f23f0f2f84dcbca1df28fd61ff6c6b"} Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.020457 4926 generic.go:334] "Generic (PLEG): container finished" podID="55cf4e17-622b-4238-894c-7b3362839762" containerID="fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71" exitCode=0 Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.020514 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerDied","Data":"fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71"} Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.021091 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tcmdr" event={"ID":"55cf4e17-622b-4238-894c-7b3362839762","Type":"ContainerDied","Data":"410e46ce45266faa35eadca767ec1291556274bee9f84a8ccde45d0d027f4f79"} Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.021118 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="410e46ce45266faa35eadca767ec1291556274bee9f84a8ccde45d0d027f4f79" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.141303 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.165783 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jscf7"] Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.168082 4926 scope.go:117] "RemoveContainer" containerID="82ce0166146e1bbf8ebbc3da9fb34d02030a47ab533510a8c51945287c47c173" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.179180 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jscf7"] Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.188675 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jdm9g"] Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.201590 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jdm9g"] Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.247591 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jnvvs\" (UniqueName: \"kubernetes.io/projected/55cf4e17-622b-4238-894c-7b3362839762-kube-api-access-jnvvs\") pod \"55cf4e17-622b-4238-894c-7b3362839762\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.247710 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-utilities\") pod \"55cf4e17-622b-4238-894c-7b3362839762\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.247750 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-catalog-content\") pod \"55cf4e17-622b-4238-894c-7b3362839762\" (UID: \"55cf4e17-622b-4238-894c-7b3362839762\") " Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.251549 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-utilities" (OuterVolumeSpecName: "utilities") pod "55cf4e17-622b-4238-894c-7b3362839762" (UID: "55cf4e17-622b-4238-894c-7b3362839762"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.290102 4926 scope.go:117] "RemoveContainer" containerID="1429a5922b3a590f2bf1144f8f98f3927e78b24d246ab9a712a81deb009e9a83" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.290340 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55cf4e17-622b-4238-894c-7b3362839762-kube-api-access-jnvvs" (OuterVolumeSpecName: "kube-api-access-jnvvs") pod "55cf4e17-622b-4238-894c-7b3362839762" (UID: "55cf4e17-622b-4238-894c-7b3362839762"). InnerVolumeSpecName "kube-api-access-jnvvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.350224 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jnvvs\" (UniqueName: \"kubernetes.io/projected/55cf4e17-622b-4238-894c-7b3362839762-kube-api-access-jnvvs\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.350250 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.361482 4926 scope.go:117] "RemoveContainer" containerID="8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.377619 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55cf4e17-622b-4238-894c-7b3362839762" (UID: "55cf4e17-622b-4238-894c-7b3362839762"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.397127 4926 scope.go:117] "RemoveContainer" containerID="6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.435090 4926 scope.go:117] "RemoveContainer" containerID="91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.453002 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55cf4e17-622b-4238-894c-7b3362839762-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.458695 4926 scope.go:117] "RemoveContainer" containerID="8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9" Nov 25 10:24:27 crc kubenswrapper[4926]: E1125 10:24:27.459153 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9\": container with ID starting with 8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9 not found: ID does not exist" containerID="8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.459194 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9"} err="failed to get container status \"8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9\": rpc error: code = NotFound desc = could not find container \"8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9\": container with ID starting with 8fc71d58c7f254ce64a0039d12679633cd09dbc00a3a8174de8f9f3f960511b9 not found: ID does not exist" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.459221 4926 scope.go:117] "RemoveContainer" containerID="6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c" Nov 25 10:24:27 crc kubenswrapper[4926]: E1125 10:24:27.459537 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c\": container with ID starting with 6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c not found: ID does not exist" containerID="6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.459561 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c"} err="failed to get container status \"6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c\": rpc error: code = NotFound desc = could not find container \"6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c\": container with ID starting with 6071a4dc6a880f5e6fa1e3274cbbed37c1282a12558e78a80c193e69e683526c not found: ID does not exist" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.459593 4926 scope.go:117] "RemoveContainer" containerID="91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385" Nov 25 10:24:27 crc kubenswrapper[4926]: E1125 10:24:27.460792 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385\": container with ID starting with 91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385 not found: ID does not exist" containerID="91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.460813 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385"} err="failed to get container status \"91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385\": rpc error: code = NotFound desc = could not find container \"91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385\": container with ID starting with 91aed678c3f31d587ff56d23de9a9fb62f8aacc6e56a316ffbea3511c0fa0385 not found: ID does not exist" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.487950 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.554112 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fg4g6\" (UniqueName: \"kubernetes.io/projected/d043df18-3a9d-4803-b6b5-398a6e0a578b-kube-api-access-fg4g6\") pod \"d043df18-3a9d-4803-b6b5-398a6e0a578b\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.554290 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-utilities\") pod \"d043df18-3a9d-4803-b6b5-398a6e0a578b\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.554331 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-catalog-content\") pod \"d043df18-3a9d-4803-b6b5-398a6e0a578b\" (UID: \"d043df18-3a9d-4803-b6b5-398a6e0a578b\") " Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.555381 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-utilities" (OuterVolumeSpecName: "utilities") pod "d043df18-3a9d-4803-b6b5-398a6e0a578b" (UID: "d043df18-3a9d-4803-b6b5-398a6e0a578b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.558870 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d043df18-3a9d-4803-b6b5-398a6e0a578b-kube-api-access-fg4g6" (OuterVolumeSpecName: "kube-api-access-fg4g6") pod "d043df18-3a9d-4803-b6b5-398a6e0a578b" (UID: "d043df18-3a9d-4803-b6b5-398a6e0a578b"). InnerVolumeSpecName "kube-api-access-fg4g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.602015 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d043df18-3a9d-4803-b6b5-398a6e0a578b" (UID: "d043df18-3a9d-4803-b6b5-398a6e0a578b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.656238 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.656283 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d043df18-3a9d-4803-b6b5-398a6e0a578b-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:27 crc kubenswrapper[4926]: I1125 10:24:27.656299 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fg4g6\" (UniqueName: \"kubernetes.io/projected/d043df18-3a9d-4803-b6b5-398a6e0a578b-kube-api-access-fg4g6\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.024373 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" path="/var/lib/kubelet/pods/1934d666-ff68-48e3-8c8e-1dbbcb39cf26/volumes" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.025110 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" path="/var/lib/kubelet/pods/1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf/volumes" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.025791 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="675b802f-5459-4178-a8b1-ad249451a47b" path="/var/lib/kubelet/pods/675b802f-5459-4178-a8b1-ad249451a47b/volumes" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.027056 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="705621a2-d616-487e-b80a-187d8d6214c2" path="/var/lib/kubelet/pods/705621a2-d616-487e-b80a-187d8d6214c2/volumes" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.027752 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2275402-9d74-4709-9908-add9b0b816c4" path="/var/lib/kubelet/pods/a2275402-9d74-4709-9908-add9b0b816c4/volumes" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.039474 4926 generic.go:334] "Generic (PLEG): container finished" podID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerID="bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b" exitCode=0 Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.039538 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vzjll" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.039572 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerDied","Data":"bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b"} Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.039628 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vzjll" event={"ID":"d043df18-3a9d-4803-b6b5-398a6e0a578b","Type":"ContainerDied","Data":"c8a7f5768a619d295478b8d50df6dd85290efab104dc974e042b1ecc3292728b"} Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.039545 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tcmdr" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.039648 4926 scope.go:117] "RemoveContainer" containerID="bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.077729 4926 scope.go:117] "RemoveContainer" containerID="43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.103611 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-tcmdr"] Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.110130 4926 scope.go:117] "RemoveContainer" containerID="0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.112619 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-tcmdr"] Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.124347 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-vzjll"] Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.137707 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-vzjll"] Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.147054 4926 scope.go:117] "RemoveContainer" containerID="bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.147543 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b\": container with ID starting with bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b not found: ID does not exist" containerID="bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.147604 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b"} err="failed to get container status \"bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b\": rpc error: code = NotFound desc = could not find container \"bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b\": container with ID starting with bd712fe8c840809534f49f879c97d8c1af66e838af59ae9ce765e47bede7024b not found: ID does not exist" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.147640 4926 scope.go:117] "RemoveContainer" containerID="43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.147994 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff\": container with ID starting with 43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff not found: ID does not exist" containerID="43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.148023 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff"} err="failed to get container status \"43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff\": rpc error: code = NotFound desc = could not find container \"43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff\": container with ID starting with 43ba0a949579667b252ccf4fcf244359459a09470f3ca9feb5890def79bd10ff not found: ID does not exist" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.148045 4926 scope.go:117] "RemoveContainer" containerID="0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.148337 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e\": container with ID starting with 0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e not found: ID does not exist" containerID="0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.148369 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e"} err="failed to get container status \"0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e\": rpc error: code = NotFound desc = could not find container \"0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e\": container with ID starting with 0d49e132ff35d0dc8027748a6d93db50d2df1a4202117e81bb452d7e82e2999e not found: ID does not exist" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.363476 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8594cdbfd6-wcgzz_a2920249-1629-4833-a8a0-0b5fee1c2303/barbican-api/0.log" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.527293 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-8594cdbfd6-wcgzz_a2920249-1629-4833-a8a0-0b5fee1c2303/barbican-api-log/0.log" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.559721 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5dd564f876-448mb_a33145e6-61bd-4706-b9cf-bf2e59906a6b/barbican-keystone-listener/0.log" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.667843 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5dd564f876-448mb_a33145e6-61bd-4706-b9cf-bf2e59906a6b/barbican-keystone-listener-log/0.log" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.780643 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-75b9f64b55-mrc6w_b069a045-2f4e-44f9-8d85-cdacf9bb1ccd/barbican-worker/0.log" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.786873 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jbn2v"] Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787281 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787301 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787314 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787321 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787333 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787340 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787352 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787358 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787369 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787375 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787384 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787390 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787396 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787402 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787408 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787416 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787429 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787437 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787445 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787452 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787465 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787471 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787483 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787490 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787502 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787508 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787524 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787533 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787542 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" containerName="container-00" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787550 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" containerName="container-00" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787563 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787570 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787587 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787593 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787608 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787615 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787631 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787638 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787651 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787660 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787672 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787680 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787696 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787702 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787715 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787720 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787730 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787736 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787745 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787751 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787760 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787766 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787776 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787782 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="extract-content" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787791 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787796 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="extract-utilities" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787802 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787808 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787817 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787823 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: E1125 10:24:28.787832 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.787837 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788091 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c1c0f63-5869-4abe-8b4a-fa9f8b4f6baf" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788107 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2275402-9d74-4709-9908-add9b0b816c4" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788117 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="25b0e9bd-d961-4465-9496-14420a9daf2c" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788132 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8e47d70-d73c-43fe-9a70-7a82c73642c0" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788146 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788158 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="55cf4e17-622b-4238-894c-7b3362839762" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788168 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="1934d666-ff68-48e3-8c8e-1dbbcb39cf26" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788177 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eab59112-b641-43af-a91b-e7c0e58ff417" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788190 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="705621a2-d616-487e-b80a-187d8d6214c2" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788196 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="675b802f-5459-4178-a8b1-ad249451a47b" containerName="registry-server" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.788208 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae4adb9a-0c49-4c55-ba5c-1483c563c1d0" containerName="container-00" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.789490 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.813818 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbn2v"] Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.854364 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-75b9f64b55-mrc6w_b069a045-2f4e-44f9-8d85-cdacf9bb1ccd/barbican-worker-log/0.log" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.878722 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-utilities\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.879159 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-catalog-content\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.879232 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5pml\" (UniqueName: \"kubernetes.io/projected/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-kube-api-access-l5pml\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.980636 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-utilities\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.980813 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-catalog-content\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.980859 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5pml\" (UniqueName: \"kubernetes.io/projected/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-kube-api-access-l5pml\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.981122 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-utilities\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:28 crc kubenswrapper[4926]: I1125 10:24:28.981270 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-catalog-content\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.013235 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5pml\" (UniqueName: \"kubernetes.io/projected/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-kube-api-access-l5pml\") pod \"community-operators-jbn2v\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.074459 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-q6g9l_ac1288a0-8cc2-49b5-b9d4-f0402c00aba9/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.105789 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a649a79-6eff-402f-8ea0-59b8649155e3/ceilometer-central-agent/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.115147 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.200607 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a649a79-6eff-402f-8ea0-59b8649155e3/ceilometer-notification-agent/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.325308 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a649a79-6eff-402f-8ea0-59b8649155e3/proxy-httpd/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.384355 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_0a649a79-6eff-402f-8ea0-59b8649155e3/sg-core/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.514858 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-xqc4v_74f34995-35ea-4875-9a69-61196081a80f/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.577634 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jbn2v"] Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.595053 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_87b02209-e235-4ce3-8f41-39fd50fd81c9/cinder-api/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.637854 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_87b02209-e235-4ce3-8f41-39fd50fd81c9/cinder-api-log/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.777179 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5d2781eb-b45c-4445-9f30-009ab5de333f/cinder-scheduler/0.log" Nov 25 10:24:29 crc kubenswrapper[4926]: I1125 10:24:29.829500 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_5d2781eb-b45c-4445-9f30-009ab5de333f/probe/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.047680 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55cf4e17-622b-4238-894c-7b3362839762" path="/var/lib/kubelet/pods/55cf4e17-622b-4238-894c-7b3362839762/volumes" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.048730 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d043df18-3a9d-4803-b6b5-398a6e0a578b" path="/var/lib/kubelet/pods/d043df18-3a9d-4803-b6b5-398a6e0a578b/volumes" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.053525 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-jsngf_0e9ec79e-2dba-4832-baa4-8b630738f623/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.064947 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerID="bc409fd5be6fb5a5df1f06884f52b593aa988279ec86ff7b73c5161da725c876" exitCode=0 Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.064994 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbn2v" event={"ID":"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4","Type":"ContainerDied","Data":"bc409fd5be6fb5a5df1f06884f52b593aa988279ec86ff7b73c5161da725c876"} Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.065021 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbn2v" event={"ID":"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4","Type":"ContainerStarted","Data":"a25e7e61f8fbcb267fa71060a15b7d71b1fde8094971fe288fdc4a2e0dfe8bd1"} Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.091985 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5tcb8_0907de16-2f17-473e-83af-94baeb44725c/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.268400 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-fbc59fbb7-nf2mz_9252a64e-adab-4db0-aa4e-6bcbb0b9458b/init/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.602453 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-fbc59fbb7-nf2mz_9252a64e-adab-4db0-aa4e-6bcbb0b9458b/init/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.624549 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-skbf5_311b8250-d497-4cf0-9b38-72e00d16f422/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.635497 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-fbc59fbb7-nf2mz_9252a64e-adab-4db0-aa4e-6bcbb0b9458b/dnsmasq-dns/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.906286 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-5c5477c6d9-t4dvb_92634e60-5ef6-4047-ad6d-0d5a58f77d05/keystone-api/0.log" Nov 25 10:24:30 crc kubenswrapper[4926]: I1125 10:24:30.910070 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2f8efeb7-cb75-4a17-bf87-85431d9e395e/kube-state-metrics/3.log" Nov 25 10:24:31 crc kubenswrapper[4926]: I1125 10:24:31.066205 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_2f8efeb7-cb75-4a17-bf87-85431d9e395e/kube-state-metrics/2.log" Nov 25 10:24:31 crc kubenswrapper[4926]: I1125 10:24:31.289516 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cdc678f6f-htchg_331a3347-d9a9-4a86-a552-58ab450406d4/neutron-api/0.log" Nov 25 10:24:31 crc kubenswrapper[4926]: I1125 10:24:31.393925 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-6cdc678f6f-htchg_331a3347-d9a9-4a86-a552-58ab450406d4/neutron-httpd/0.log" Nov 25 10:24:31 crc kubenswrapper[4926]: I1125 10:24:31.802742 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_de6e1390-0172-4a3f-b6cb-5abb2a5633b2/nova-api-log/0.log" Nov 25 10:24:31 crc kubenswrapper[4926]: I1125 10:24:31.803508 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_de6e1390-0172-4a3f-b6cb-5abb2a5633b2/nova-api-api/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.064001 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_5bbaac50-a726-4ae4-81b9-cb19e7d2d81a/nova-cell0-conductor-conductor/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.092397 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerID="d8c4476fa96f97a2a65099621b57a01822397c6be4e837b2ef48f2caa71ead13" exitCode=0 Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.092437 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbn2v" event={"ID":"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4","Type":"ContainerDied","Data":"d8c4476fa96f97a2a65099621b57a01822397c6be4e837b2ef48f2caa71ead13"} Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.175120 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_05c96c75-c327-4561-ac40-59800d92b9a6/nova-cell1-conductor-conductor/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.426582 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_ff3b4281-7815-4363-91b6-fabe18bfdb70/nova-cell1-novncproxy-novncproxy/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.474842 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e1abc2c8-9361-47b7-88f1-818915a5800e/nova-metadata-log/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.846408 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_d6e36e84-7af6-4a0b-98c8-dc100ece345b/nova-scheduler-scheduler/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.910358 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_156bc1bd-066f-4071-8d4a-b2124fe381bb/mysql-bootstrap/0.log" Nov 25 10:24:32 crc kubenswrapper[4926]: I1125 10:24:32.911595 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_e1abc2c8-9361-47b7-88f1-818915a5800e/nova-metadata-metadata/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.190984 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_156bc1bd-066f-4071-8d4a-b2124fe381bb/galera/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.217593 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2a7b065a-b959-4809-8f8d-1a36b09579c9/mysql-bootstrap/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.235222 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_156bc1bd-066f-4071-8d4a-b2124fe381bb/mysql-bootstrap/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.411368 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2a7b065a-b959-4809-8f8d-1a36b09579c9/mysql-bootstrap/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.584510 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2a7b065a-b959-4809-8f8d-1a36b09579c9/galera/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.589746 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_02f6bba3-15f5-40ad-9624-deb63a912775/openstackclient/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.893090 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-5xppx_1a1386c3-686d-4f5f-85b1-2f721ed3be80/openstack-network-exporter/0.log" Nov 25 10:24:33 crc kubenswrapper[4926]: I1125 10:24:33.952420 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-6dkhn_af231b27-f9dd-466d-83d0-05484183099f/ovn-controller/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.118384 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbn2v" event={"ID":"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4","Type":"ContainerStarted","Data":"72e1e98c344cff688fc381c912d7e4900ff38edb51aa91bc12dde51099ea2c16"} Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.143306 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jbn2v" podStartSLOduration=3.247265756 podStartE2EDuration="6.143287486s" podCreationTimestamp="2025-11-25 10:24:28 +0000 UTC" firstStartedPulling="2025-11-25 10:24:30.066613009 +0000 UTC m=+2166.750569436" lastFinishedPulling="2025-11-25 10:24:32.962634739 +0000 UTC m=+2169.646591166" observedRunningTime="2025-11-25 10:24:34.140715212 +0000 UTC m=+2170.824671639" watchObservedRunningTime="2025-11-25 10:24:34.143287486 +0000 UTC m=+2170.827243913" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.227976 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkcjx_dbdb45ad-32a5-4edc-aaa3-d907b5272b8a/ovsdb-server-init/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.408734 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkcjx_dbdb45ad-32a5-4edc-aaa3-d907b5272b8a/ovsdb-server/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.413908 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkcjx_dbdb45ad-32a5-4edc-aaa3-d907b5272b8a/ovsdb-server-init/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.444524 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-dkcjx_dbdb45ad-32a5-4edc-aaa3-d907b5272b8a/ovs-vswitchd/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.721745 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_513417d4-ae93-44eb-a5e9-f16094bd3330/openstack-network-exporter/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.736612 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_513417d4-ae93-44eb-a5e9-f16094bd3330/ovn-northd/0.log" Nov 25 10:24:34 crc kubenswrapper[4926]: I1125 10:24:34.871406 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_17396f6f-d876-49fe-9294-fe074f9da844/openstack-network-exporter/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.006364 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_17396f6f-d876-49fe-9294-fe074f9da844/ovsdbserver-nb/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.019712 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e7753413-a221-4236-9003-e722280e15bf/openstack-network-exporter/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.209162 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_e7753413-a221-4236-9003-e722280e15bf/ovsdbserver-sb/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.320386 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d785fcd8-5cqwz_e84974e8-06ae-4d64-8a23-5054daa2a0e8/placement-log/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.331679 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-d785fcd8-5cqwz_e84974e8-06ae-4d64-8a23-5054daa2a0e8/placement-api/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.518743 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_30c23e52-ddbc-4e38-85f6-6a7b32e67d12/setup-container/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.837746 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4bccf53f-5bb6-4e30-876e-e3d6be02f9fe/setup-container/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.886790 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_30c23e52-ddbc-4e38-85f6-6a7b32e67d12/rabbitmq/0.log" Nov 25 10:24:35 crc kubenswrapper[4926]: I1125 10:24:35.904929 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_30c23e52-ddbc-4e38-85f6-6a7b32e67d12/setup-container/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.125420 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4bccf53f-5bb6-4e30-876e-e3d6be02f9fe/setup-container/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.182307 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_4bccf53f-5bb6-4e30-876e-e3d6be02f9fe/rabbitmq/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.211726 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fnp9t_c2514ec8-a2ab-485d-9385-838ac88416d0/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.426035 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-4jbkp_ae06cbe1-66b6-4d10-b2fe-99834e3fc42c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.532102 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-dt7h7_f04c96be-4b0e-43a9-9362-90c38248ac72/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.680218 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-5fxd9_6f0df4f6-baeb-4c84-baa9-551970e17317/ssh-known-hosts-edpm-deployment/0.log" Nov 25 10:24:36 crc kubenswrapper[4926]: I1125 10:24:36.820040 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-lkln8_2f83a69b-6808-4409-b29a-72c78653c714/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Nov 25 10:24:38 crc kubenswrapper[4926]: I1125 10:24:38.042737 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_dd48201c-f958-4647-afe6-055c685cb255/memcached/0.log" Nov 25 10:24:39 crc kubenswrapper[4926]: I1125 10:24:39.115263 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:39 crc kubenswrapper[4926]: I1125 10:24:39.116660 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:39 crc kubenswrapper[4926]: I1125 10:24:39.169340 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:40 crc kubenswrapper[4926]: I1125 10:24:40.266363 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:40 crc kubenswrapper[4926]: I1125 10:24:40.309263 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbn2v"] Nov 25 10:24:42 crc kubenswrapper[4926]: I1125 10:24:42.217677 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jbn2v" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="registry-server" containerID="cri-o://72e1e98c344cff688fc381c912d7e4900ff38edb51aa91bc12dde51099ea2c16" gracePeriod=2 Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.236259 4926 generic.go:334] "Generic (PLEG): container finished" podID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerID="72e1e98c344cff688fc381c912d7e4900ff38edb51aa91bc12dde51099ea2c16" exitCode=0 Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.236594 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbn2v" event={"ID":"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4","Type":"ContainerDied","Data":"72e1e98c344cff688fc381c912d7e4900ff38edb51aa91bc12dde51099ea2c16"} Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.483112 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.565149 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-utilities\") pod \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.565566 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5pml\" (UniqueName: \"kubernetes.io/projected/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-kube-api-access-l5pml\") pod \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.565680 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-catalog-content\") pod \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\" (UID: \"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4\") " Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.566199 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-utilities" (OuterVolumeSpecName: "utilities") pod "2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" (UID: "2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.572910 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-kube-api-access-l5pml" (OuterVolumeSpecName: "kube-api-access-l5pml") pod "2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" (UID: "2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4"). InnerVolumeSpecName "kube-api-access-l5pml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.628979 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" (UID: "2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.667388 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5pml\" (UniqueName: \"kubernetes.io/projected/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-kube-api-access-l5pml\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.667426 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:43 crc kubenswrapper[4926]: I1125 10:24:43.667435 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.019212 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.019320 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.249826 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jbn2v" event={"ID":"2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4","Type":"ContainerDied","Data":"a25e7e61f8fbcb267fa71060a15b7d71b1fde8094971fe288fdc4a2e0dfe8bd1"} Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.249884 4926 scope.go:117] "RemoveContainer" containerID="72e1e98c344cff688fc381c912d7e4900ff38edb51aa91bc12dde51099ea2c16" Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.249900 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jbn2v" Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.278784 4926 scope.go:117] "RemoveContainer" containerID="d8c4476fa96f97a2a65099621b57a01822397c6be4e837b2ef48f2caa71ead13" Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.279151 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jbn2v"] Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.289111 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jbn2v"] Nov 25 10:24:44 crc kubenswrapper[4926]: I1125 10:24:44.301473 4926 scope.go:117] "RemoveContainer" containerID="bc409fd5be6fb5a5df1f06884f52b593aa988279ec86ff7b73c5161da725c876" Nov 25 10:24:46 crc kubenswrapper[4926]: I1125 10:24:46.022482 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" path="/var/lib/kubelet/pods/2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4/volumes" Nov 25 10:24:56 crc kubenswrapper[4926]: I1125 10:24:56.942059 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/util/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.140273 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/util/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.166639 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/pull/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.178254 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/pull/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.355648 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/util/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.416790 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/extract/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.435191 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_97921fde98405fcf786bdfad979ee1b6baaacc76ca58626fa30c19a39anpfk2_aa8e7b73-9640-4c71-9615-41c32b22f48c/pull/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.613454 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-fr2mr_194893a3-1c67-4a28-a67e-7d7eeef304a7/kube-rbac-proxy/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.667197 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-fr2mr_194893a3-1c67-4a28-a67e-7d7eeef304a7/manager/1.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.759467 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-86dc4d89c8-fr2mr_194893a3-1c67-4a28-a67e-7d7eeef304a7/manager/2.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.811464 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-67llk_09b3e2a6-3249-4ca3-afc9-39dd744ddd10/kube-rbac-proxy/0.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.851126 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-67llk_09b3e2a6-3249-4ca3-afc9-39dd744ddd10/manager/2.log" Nov 25 10:24:57 crc kubenswrapper[4926]: I1125 10:24:57.956797 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-79856dc55c-67llk_09b3e2a6-3249-4ca3-afc9-39dd744ddd10/manager/1.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.063391 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-4wxmj_8292ff2d-08ba-4be8-a268-5c8b6d639087/kube-rbac-proxy/0.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.068730 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-4wxmj_8292ff2d-08ba-4be8-a268-5c8b6d639087/manager/2.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.165264 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-7d695c9b56-4wxmj_8292ff2d-08ba-4be8-a268-5c8b6d639087/manager/1.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.272594 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-75fd7f55fb-p7v2x_5249ee3d-0121-42bf-bc4a-7d05a8410a9f/kube-rbac-proxy/0.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.285466 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-75fd7f55fb-p7v2x_5249ee3d-0121-42bf-bc4a-7d05a8410a9f/manager/2.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.348510 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-75fd7f55fb-p7v2x_5249ee3d-0121-42bf-bc4a-7d05a8410a9f/manager/1.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.468218 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-4d4g5_df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a/manager/2.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.469247 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-4d4g5_df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a/kube-rbac-proxy/0.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.554642 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-774b86978c-4d4g5_df9a8fa9-e7e8-42c5-8d87-adcbccd1a06a/manager/1.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.665684 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-kqv6n_1154878f-e8a5-42de-916f-73276b979e74/kube-rbac-proxy/0.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.713691 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-kqv6n_1154878f-e8a5-42de-916f-73276b979e74/manager/2.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.751166 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c9694994-kqv6n_1154878f-e8a5-42de-916f-73276b979e74/manager/1.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.868539 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858778c9dc-lwgz6_b6772bac-4625-41d7-915d-2fdabd9916b7/kube-rbac-proxy/0.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.921282 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858778c9dc-lwgz6_b6772bac-4625-41d7-915d-2fdabd9916b7/manager/2.log" Nov 25 10:24:58 crc kubenswrapper[4926]: I1125 10:24:58.933729 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-858778c9dc-lwgz6_b6772bac-4625-41d7-915d-2fdabd9916b7/manager/1.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.060415 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-wt256_0ed1bf71-b826-4854-873d-7c6fcf992bbd/kube-rbac-proxy/0.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.107485 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-wt256_0ed1bf71-b826-4854-873d-7c6fcf992bbd/manager/1.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.109678 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-5bfcdc958c-wt256_0ed1bf71-b826-4854-873d-7c6fcf992bbd/manager/2.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.287529 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-82vvs_bc152fdc-1bab-4528-992d-0883671d7372/kube-rbac-proxy/0.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.417342 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-82vvs_bc152fdc-1bab-4528-992d-0883671d7372/manager/1.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.421664 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-748dc6576f-82vvs_bc152fdc-1bab-4528-992d-0883671d7372/manager/2.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.584153 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-2b444_409d9035-812c-4b37-b495-d6554315fb03/manager/2.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.591502 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-2b444_409d9035-812c-4b37-b495-d6554315fb03/kube-rbac-proxy/0.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.607644 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-58bb8d67cc-2b444_409d9035-812c-4b37-b495-d6554315fb03/manager/1.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.777938 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-hgcnt_5bf64213-939f-4f49-9030-9f9dcc78f8bb/kube-rbac-proxy/0.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.802227 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-hgcnt_5bf64213-939f-4f49-9030-9f9dcc78f8bb/manager/1.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.853581 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-cb6c4fdb7-hgcnt_5bf64213-939f-4f49-9030-9f9dcc78f8bb/manager/2.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.953828 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-947wd_dd863871-17e8-4a97-bc8b-5dd02f807225/kube-rbac-proxy/0.log" Nov 25 10:24:59 crc kubenswrapper[4926]: I1125 10:24:59.981472 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-947wd_dd863871-17e8-4a97-bc8b-5dd02f807225/manager/2.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.043471 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-7c57c8bbc4-947wd_dd863871-17e8-4a97-bc8b-5dd02f807225/manager/1.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.170873 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tmxx7_fa24e0a8-f9a4-4160-bac5-4d73e8579fd8/manager/2.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.188560 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tmxx7_fa24e0a8-f9a4-4160-bac5-4d73e8579fd8/kube-rbac-proxy/0.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.238828 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-79556f57fc-tmxx7_fa24e0a8-f9a4-4160-bac5-4d73e8579fd8/manager/1.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.343881 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-mt4w9_f74db279-9026-4869-a147-d45bf581b618/kube-rbac-proxy/0.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.382708 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-mt4w9_f74db279-9026-4869-a147-d45bf581b618/manager/2.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.405875 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-fd75fd47d-mt4w9_f74db279-9026-4869-a147-d45bf581b618/manager/1.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.559389 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs_db149e24-86d6-4f34-abb0-82229e25e9d7/manager/1.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.585011 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs_db149e24-86d6-4f34-abb0-82229e25e9d7/kube-rbac-proxy/0.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.645144 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-544b9bb9-vwsgs_db149e24-86d6-4f34-abb0-82229e25e9d7/manager/0.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.783950 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86896d4c4b-ssjxq_4403c995-1767-4968-b0f4-0a4f0d70a188/manager/2.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.814533 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-86896d4c4b-ssjxq_4403c995-1767-4968-b0f4-0a4f0d70a188/manager/1.log" Nov 25 10:25:00 crc kubenswrapper[4926]: I1125 10:25:00.875924 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6444d758f4-lgw8g_5cf84563-5163-4108-92dc-1dbf757212de/operator/1.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.091122 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-h4hql_047b7e2a-f92d-41be-aa69-f3807ae9b0c4/registry-server/0.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.091502 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-6444d758f4-lgw8g_5cf84563-5163-4108-92dc-1dbf757212de/operator/0.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.118795 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-hr2gx_edb138df-3c28-4b7d-b285-2fe43094ead3/kube-rbac-proxy/0.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.197856 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-hr2gx_edb138df-3c28-4b7d-b285-2fe43094ead3/manager/2.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.272985 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-66cf5c67ff-hr2gx_edb138df-3c28-4b7d-b285-2fe43094ead3/manager/1.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.301402 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-lwls8_ae1df4bc-d341-49fd-a49a-e02b4d8714d1/kube-rbac-proxy/0.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.377426 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-lwls8_ae1df4bc-d341-49fd-a49a-e02b4d8714d1/manager/2.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.399892 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5db546f9d9-lwls8_ae1df4bc-d341-49fd-a49a-e02b4d8714d1/manager/1.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.559811 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-zdqj6_03cc50ce-1146-4953-901e-19d5bd7c468e/operator/2.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.568270 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-zdqj6_03cc50ce-1146-4953-901e-19d5bd7c468e/operator/3.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.613483 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-kspmf_416161f0-e9fb-4909-83c0-dfe7ad15885f/kube-rbac-proxy/0.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.696516 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-kspmf_416161f0-e9fb-4909-83c0-dfe7ad15885f/manager/2.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.789334 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-6fdc4fcf86-kspmf_416161f0-e9fb-4909-83c0-dfe7ad15885f/manager/1.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.828881 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-5wdlm_7b32eddd-fe58-4854-bc45-a97ec5349f86/kube-rbac-proxy/0.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.882617 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-5wdlm_7b32eddd-fe58-4854-bc45-a97ec5349f86/manager/2.log" Nov 25 10:25:01 crc kubenswrapper[4926]: I1125 10:25:01.890346 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-567f98c9d-5wdlm_7b32eddd-fe58-4854-bc45-a97ec5349f86/manager/1.log" Nov 25 10:25:02 crc kubenswrapper[4926]: I1125 10:25:02.011695 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-fbbkn_2494b104-d923-4de0-82e0-8f4ff0e9c5f8/kube-rbac-proxy/0.log" Nov 25 10:25:02 crc kubenswrapper[4926]: I1125 10:25:02.108646 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-fbbkn_2494b104-d923-4de0-82e0-8f4ff0e9c5f8/manager/1.log" Nov 25 10:25:02 crc kubenswrapper[4926]: I1125 10:25:02.135044 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5cb74df96-fbbkn_2494b104-d923-4de0-82e0-8f4ff0e9c5f8/manager/0.log" Nov 25 10:25:02 crc kubenswrapper[4926]: I1125 10:25:02.212712 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vtrb2_d95f6ac6-9ecf-4cef-ad4e-1389a10f303e/kube-rbac-proxy/0.log" Nov 25 10:25:02 crc kubenswrapper[4926]: I1125 10:25:02.244668 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vtrb2_d95f6ac6-9ecf-4cef-ad4e-1389a10f303e/manager/2.log" Nov 25 10:25:02 crc kubenswrapper[4926]: I1125 10:25:02.305428 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-864885998-vtrb2_d95f6ac6-9ecf-4cef-ad4e-1389a10f303e/manager/1.log" Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.018796 4926 patch_prober.go:28] interesting pod/machine-config-daemon-b82dg container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.019315 4926 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.025615 4926 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.026386 4926 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28"} pod="openshift-machine-config-operator/machine-config-daemon-b82dg" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.026457 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerName="machine-config-daemon" containerID="cri-o://eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" gracePeriod=600 Nov 25 10:25:14 crc kubenswrapper[4926]: E1125 10:25:14.145032 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.494467 4926 generic.go:334] "Generic (PLEG): container finished" podID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" exitCode=0 Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.494570 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerDied","Data":"eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28"} Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.494778 4926 scope.go:117] "RemoveContainer" containerID="d6f381f071189398b29183f9732862d78affd0630860a5419ed3e39c642baae0" Nov 25 10:25:14 crc kubenswrapper[4926]: I1125 10:25:14.495166 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:25:14 crc kubenswrapper[4926]: E1125 10:25:14.495415 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:25:18 crc kubenswrapper[4926]: I1125 10:25:18.439773 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-xnjf9_146577d7-d2c0-4bf4-9d6b-236967454f83/control-plane-machine-set-operator/0.log" Nov 25 10:25:18 crc kubenswrapper[4926]: I1125 10:25:18.569776 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9pvq7_ec3f0e93-5599-45c0-b6d0-5e16be0df94c/kube-rbac-proxy/0.log" Nov 25 10:25:18 crc kubenswrapper[4926]: I1125 10:25:18.628498 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-9pvq7_ec3f0e93-5599-45c0-b6d0-5e16be0df94c/machine-api-operator/0.log" Nov 25 10:25:26 crc kubenswrapper[4926]: I1125 10:25:26.012953 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:25:26 crc kubenswrapper[4926]: E1125 10:25:26.013694 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:25:30 crc kubenswrapper[4926]: I1125 10:25:30.005295 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-sxxx6_1fddcfdd-0fb9-4153-a227-b28e6faccb40/cert-manager-controller/1.log" Nov 25 10:25:30 crc kubenswrapper[4926]: I1125 10:25:30.005334 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-sxxx6_1fddcfdd-0fb9-4153-a227-b28e6faccb40/cert-manager-controller/0.log" Nov 25 10:25:30 crc kubenswrapper[4926]: I1125 10:25:30.184657 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-hpkft_96a34dc4-fd0b-43c2-bed3-741482089f7b/cert-manager-cainjector/1.log" Nov 25 10:25:30 crc kubenswrapper[4926]: I1125 10:25:30.187775 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-hpkft_96a34dc4-fd0b-43c2-bed3-741482089f7b/cert-manager-cainjector/0.log" Nov 25 10:25:30 crc kubenswrapper[4926]: I1125 10:25:30.323222 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-8pj2r_6b70e250-e5c4-4754-b5d1-a4fe76f60b00/cert-manager-webhook/0.log" Nov 25 10:25:40 crc kubenswrapper[4926]: I1125 10:25:40.012310 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:25:40 crc kubenswrapper[4926]: E1125 10:25:40.013099 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:25:41 crc kubenswrapper[4926]: I1125 10:25:41.094734 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-5874bd7bc5-t42j4_b5c62b83-eae8-455e-9c6e-166a9810c3f3/nmstate-console-plugin/0.log" Nov 25 10:25:41 crc kubenswrapper[4926]: I1125 10:25:41.265073 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-42kdg_11e4ce92-8234-4e31-a300-1896267b083c/nmstate-handler/0.log" Nov 25 10:25:41 crc kubenswrapper[4926]: I1125 10:25:41.317463 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-lbjdd_230299a9-48a6-405b-9569-33b7110107be/kube-rbac-proxy/0.log" Nov 25 10:25:41 crc kubenswrapper[4926]: I1125 10:25:41.326291 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-5dcf9c57c5-lbjdd_230299a9-48a6-405b-9569-33b7110107be/nmstate-metrics/0.log" Nov 25 10:25:41 crc kubenswrapper[4926]: I1125 10:25:41.510013 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-557fdffb88-qh5st_e766d24a-f1e5-4686-a6b3-539b906cdcfa/nmstate-operator/0.log" Nov 25 10:25:41 crc kubenswrapper[4926]: I1125 10:25:41.548531 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6b89b748d8-cbvck_a04b62fe-2ebd-4b20-b4b0-93038ead2692/nmstate-webhook/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.031291 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-bvqpc_765233dc-e1b5-47b8-8eb5-8cef0a62f428/kube-rbac-proxy/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.232510 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-bvqpc_765233dc-e1b5-47b8-8eb5-8cef0a62f428/controller/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.237757 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-frr-files/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.413188 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-frr-files/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.459153 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-reloader/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.478279 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-reloader/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.490150 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-metrics/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.626817 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-frr-files/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.647212 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-reloader/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.665766 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-metrics/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.666176 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-metrics/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.835477 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-metrics/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.838825 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-frr-files/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.841560 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/cp-reloader/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.906519 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/controller/0.log" Nov 25 10:25:54 crc kubenswrapper[4926]: I1125 10:25:54.999359 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/frr-metrics/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.012752 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:25:55 crc kubenswrapper[4926]: E1125 10:25:55.013075 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.033203 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/kube-rbac-proxy/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.093597 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/kube-rbac-proxy-frr/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.247486 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/reloader/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.283126 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-6ztkb_1a8ae06b-bbc4-4854-9ebd-a88bf4c7cde2/frr-k8s-webhook-server/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.484678 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-65f5fc5bb5-r6kqg_0cb1f5f8-6fcb-4664-8fcc-74e416d63eef/manager/2.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.497876 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-65f5fc5bb5-r6kqg_0cb1f5f8-6fcb-4664-8fcc-74e416d63eef/manager/3.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.747650 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6f64c7cdd5-r4h6s_7c5e2728-0650-42d8-b0e0-212e87536355/webhook-server/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.839358 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6rxct_7586af8a-a829-4c9f-9589-a3643c147f1d/kube-rbac-proxy/0.log" Nov 25 10:25:55 crc kubenswrapper[4926]: I1125 10:25:55.970460 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-49pz2_99b2fe2e-14bb-4dda-9152-bb593b6cb80e/frr/0.log" Nov 25 10:25:56 crc kubenswrapper[4926]: I1125 10:25:56.372529 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-6rxct_7586af8a-a829-4c9f-9589-a3643c147f1d/speaker/0.log" Nov 25 10:26:06 crc kubenswrapper[4926]: I1125 10:26:06.011985 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:26:06 crc kubenswrapper[4926]: E1125 10:26:06.012741 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:26:21 crc kubenswrapper[4926]: I1125 10:26:21.012814 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:26:21 crc kubenswrapper[4926]: E1125 10:26:21.013638 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:26:32 crc kubenswrapper[4926]: I1125 10:26:32.014612 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:26:32 crc kubenswrapper[4926]: E1125 10:26:32.015441 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:26:47 crc kubenswrapper[4926]: I1125 10:26:47.013458 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:26:47 crc kubenswrapper[4926]: E1125 10:26:47.014418 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:26:59 crc kubenswrapper[4926]: I1125 10:26:59.013200 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:26:59 crc kubenswrapper[4926]: E1125 10:26:59.014114 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:27:09 crc kubenswrapper[4926]: I1125 10:27:09.517390 4926 generic.go:334] "Generic (PLEG): container finished" podID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerID="585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22" exitCode=0 Nov 25 10:27:09 crc kubenswrapper[4926]: I1125 10:27:09.517471 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" event={"ID":"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15","Type":"ContainerDied","Data":"585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22"} Nov 25 10:27:09 crc kubenswrapper[4926]: I1125 10:27:09.518439 4926 scope.go:117] "RemoveContainer" containerID="585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22" Nov 25 10:27:09 crc kubenswrapper[4926]: I1125 10:27:09.638286 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wtrhn_must-gather-bzlp8_89fb69a6-8f3d-4f5d-b3f9-104be90c2d15/gather/0.log" Nov 25 10:27:11 crc kubenswrapper[4926]: E1125 10:27:11.253516 4926 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.90:35190->38.129.56.90:40315: write tcp 38.129.56.90:35190->38.129.56.90:40315: write: connection reset by peer Nov 25 10:27:14 crc kubenswrapper[4926]: I1125 10:27:14.019179 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:27:14 crc kubenswrapper[4926]: E1125 10:27:14.019950 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:27:16 crc kubenswrapper[4926]: I1125 10:27:16.699203 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-wtrhn/must-gather-bzlp8"] Nov 25 10:27:16 crc kubenswrapper[4926]: I1125 10:27:16.699733 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="copy" containerID="cri-o://5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602" gracePeriod=2 Nov 25 10:27:16 crc kubenswrapper[4926]: I1125 10:27:16.717770 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-wtrhn/must-gather-bzlp8"] Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.122168 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wtrhn_must-gather-bzlp8_89fb69a6-8f3d-4f5d-b3f9-104be90c2d15/copy/0.log" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.122926 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.305454 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccdm2\" (UniqueName: \"kubernetes.io/projected/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-kube-api-access-ccdm2\") pod \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.307288 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-must-gather-output\") pod \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\" (UID: \"89fb69a6-8f3d-4f5d-b3f9-104be90c2d15\") " Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.311544 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-kube-api-access-ccdm2" (OuterVolumeSpecName: "kube-api-access-ccdm2") pod "89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" (UID: "89fb69a6-8f3d-4f5d-b3f9-104be90c2d15"). InnerVolumeSpecName "kube-api-access-ccdm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.409729 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccdm2\" (UniqueName: \"kubernetes.io/projected/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-kube-api-access-ccdm2\") on node \"crc\" DevicePath \"\"" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.451320 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" (UID: "89fb69a6-8f3d-4f5d-b3f9-104be90c2d15"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.511692 4926 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.603343 4926 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-wtrhn_must-gather-bzlp8_89fb69a6-8f3d-4f5d-b3f9-104be90c2d15/copy/0.log" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.603796 4926 generic.go:334] "Generic (PLEG): container finished" podID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerID="5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602" exitCode=143 Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.603846 4926 scope.go:117] "RemoveContainer" containerID="5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.603875 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-wtrhn/must-gather-bzlp8" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.626159 4926 scope.go:117] "RemoveContainer" containerID="585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.692673 4926 scope.go:117] "RemoveContainer" containerID="5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602" Nov 25 10:27:17 crc kubenswrapper[4926]: E1125 10:27:17.693204 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602\": container with ID starting with 5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602 not found: ID does not exist" containerID="5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.693241 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602"} err="failed to get container status \"5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602\": rpc error: code = NotFound desc = could not find container \"5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602\": container with ID starting with 5d42632b900280b624d4a80e26d201ae53f07386472d16dbceaa5ddd975cb602 not found: ID does not exist" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.693266 4926 scope.go:117] "RemoveContainer" containerID="585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22" Nov 25 10:27:17 crc kubenswrapper[4926]: E1125 10:27:17.693633 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22\": container with ID starting with 585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22 not found: ID does not exist" containerID="585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22" Nov 25 10:27:17 crc kubenswrapper[4926]: I1125 10:27:17.693659 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22"} err="failed to get container status \"585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22\": rpc error: code = NotFound desc = could not find container \"585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22\": container with ID starting with 585b5f983316ec6b1a16294357cfe2d6561e414b8cd47dcc21986529e6dd0f22 not found: ID does not exist" Nov 25 10:27:18 crc kubenswrapper[4926]: I1125 10:27:18.021354 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" path="/var/lib/kubelet/pods/89fb69a6-8f3d-4f5d-b3f9-104be90c2d15/volumes" Nov 25 10:27:27 crc kubenswrapper[4926]: I1125 10:27:27.011759 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:27:27 crc kubenswrapper[4926]: E1125 10:27:27.012468 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:27:41 crc kubenswrapper[4926]: I1125 10:27:41.013311 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:27:41 crc kubenswrapper[4926]: E1125 10:27:41.014000 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:27:54 crc kubenswrapper[4926]: I1125 10:27:54.017702 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:27:54 crc kubenswrapper[4926]: E1125 10:27:54.018599 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:28:09 crc kubenswrapper[4926]: I1125 10:28:09.012955 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:28:09 crc kubenswrapper[4926]: E1125 10:28:09.013833 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:28:23 crc kubenswrapper[4926]: I1125 10:28:23.012426 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:28:23 crc kubenswrapper[4926]: E1125 10:28:23.013263 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.780054 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j9t64"] Nov 25 10:28:32 crc kubenswrapper[4926]: E1125 10:28:32.781111 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="extract-content" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781132 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="extract-content" Nov 25 10:28:32 crc kubenswrapper[4926]: E1125 10:28:32.781142 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="registry-server" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781151 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="registry-server" Nov 25 10:28:32 crc kubenswrapper[4926]: E1125 10:28:32.781176 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="gather" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781183 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="gather" Nov 25 10:28:32 crc kubenswrapper[4926]: E1125 10:28:32.781209 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="extract-utilities" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781217 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="extract-utilities" Nov 25 10:28:32 crc kubenswrapper[4926]: E1125 10:28:32.781258 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="copy" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781265 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="copy" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781419 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e5b4a82-4fa1-4e17-9b7d-f91525eff4c4" containerName="registry-server" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781441 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="copy" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.781455 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="89fb69a6-8f3d-4f5d-b3f9-104be90c2d15" containerName="gather" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.783023 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.790150 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9t64"] Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.895024 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-catalog-content\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.895100 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-utilities\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.895403 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlzj6\" (UniqueName: \"kubernetes.io/projected/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-kube-api-access-vlzj6\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.997285 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-catalog-content\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.997329 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-utilities\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.997418 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlzj6\" (UniqueName: \"kubernetes.io/projected/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-kube-api-access-vlzj6\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.997883 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-utilities\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:32 crc kubenswrapper[4926]: I1125 10:28:32.997891 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-catalog-content\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:33 crc kubenswrapper[4926]: I1125 10:28:33.022458 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlzj6\" (UniqueName: \"kubernetes.io/projected/b265e6d7-6452-4cf7-8d79-9ac0bd092a3f-kube-api-access-vlzj6\") pod \"redhat-operators-j9t64\" (UID: \"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f\") " pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:33 crc kubenswrapper[4926]: I1125 10:28:33.133638 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:33 crc kubenswrapper[4926]: I1125 10:28:33.597602 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9t64"] Nov 25 10:28:34 crc kubenswrapper[4926]: I1125 10:28:34.020432 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:28:34 crc kubenswrapper[4926]: E1125 10:28:34.022321 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:28:34 crc kubenswrapper[4926]: I1125 10:28:34.300491 4926 generic.go:334] "Generic (PLEG): container finished" podID="b265e6d7-6452-4cf7-8d79-9ac0bd092a3f" containerID="063821cc9f74f9d53cbe75f2cfdae6399e12b8922550cb1bb47423b8eb0c56b4" exitCode=0 Nov 25 10:28:34 crc kubenswrapper[4926]: I1125 10:28:34.300537 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9t64" event={"ID":"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f","Type":"ContainerDied","Data":"063821cc9f74f9d53cbe75f2cfdae6399e12b8922550cb1bb47423b8eb0c56b4"} Nov 25 10:28:34 crc kubenswrapper[4926]: I1125 10:28:34.300564 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9t64" event={"ID":"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f","Type":"ContainerStarted","Data":"e58e302631fcb3283e5b5ed851bbc4b504a2391c8eabd0af05dcc3b8ff704b48"} Nov 25 10:28:34 crc kubenswrapper[4926]: I1125 10:28:34.302533 4926 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 25 10:28:42 crc kubenswrapper[4926]: I1125 10:28:42.374744 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9t64" event={"ID":"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f","Type":"ContainerStarted","Data":"c3b6780dfe323c195c1598819aeee390f7728eed889e7bc9002ae18a6df0f554"} Nov 25 10:28:43 crc kubenswrapper[4926]: I1125 10:28:43.387536 4926 generic.go:334] "Generic (PLEG): container finished" podID="b265e6d7-6452-4cf7-8d79-9ac0bd092a3f" containerID="c3b6780dfe323c195c1598819aeee390f7728eed889e7bc9002ae18a6df0f554" exitCode=0 Nov 25 10:28:43 crc kubenswrapper[4926]: I1125 10:28:43.387601 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9t64" event={"ID":"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f","Type":"ContainerDied","Data":"c3b6780dfe323c195c1598819aeee390f7728eed889e7bc9002ae18a6df0f554"} Nov 25 10:28:44 crc kubenswrapper[4926]: I1125 10:28:44.400856 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j9t64" event={"ID":"b265e6d7-6452-4cf7-8d79-9ac0bd092a3f","Type":"ContainerStarted","Data":"764bd258b9a53eeacf4b7758cfb0f743d29d91886de198961a09b325d666c79d"} Nov 25 10:28:44 crc kubenswrapper[4926]: I1125 10:28:44.436347 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j9t64" podStartSLOduration=2.892091563 podStartE2EDuration="12.436320202s" podCreationTimestamp="2025-11-25 10:28:32 +0000 UTC" firstStartedPulling="2025-11-25 10:28:34.302313487 +0000 UTC m=+2410.986269914" lastFinishedPulling="2025-11-25 10:28:43.846542116 +0000 UTC m=+2420.530498553" observedRunningTime="2025-11-25 10:28:44.431250625 +0000 UTC m=+2421.115207052" watchObservedRunningTime="2025-11-25 10:28:44.436320202 +0000 UTC m=+2421.120276629" Nov 25 10:28:48 crc kubenswrapper[4926]: I1125 10:28:48.021129 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:28:48 crc kubenswrapper[4926]: E1125 10:28:48.022074 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.134111 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.135521 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.180431 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.527309 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j9t64" Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.636261 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j9t64"] Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.684824 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9r4r6"] Nov 25 10:28:53 crc kubenswrapper[4926]: I1125 10:28:53.685179 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-9r4r6" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="registry-server" containerID="cri-o://3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96" gracePeriod=2 Nov 25 10:28:56 crc kubenswrapper[4926]: E1125 10:28:56.162129 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96 is running failed: container process not found" containerID="3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 10:28:56 crc kubenswrapper[4926]: E1125 10:28:56.163855 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96 is running failed: container process not found" containerID="3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 10:28:56 crc kubenswrapper[4926]: E1125 10:28:56.164239 4926 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96 is running failed: container process not found" containerID="3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96" cmd=["grpc_health_probe","-addr=:50051"] Nov 25 10:28:56 crc kubenswrapper[4926]: E1125 10:28:56.164275 4926 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-9r4r6" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="registry-server" Nov 25 10:28:57 crc kubenswrapper[4926]: I1125 10:28:57.518360 4926 generic.go:334] "Generic (PLEG): container finished" podID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerID="3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96" exitCode=0 Nov 25 10:28:57 crc kubenswrapper[4926]: I1125 10:28:57.518685 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerDied","Data":"3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96"} Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.679630 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.802656 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-catalog-content\") pod \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.802805 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b74zs\" (UniqueName: \"kubernetes.io/projected/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-kube-api-access-b74zs\") pod \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.802861 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-utilities\") pod \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\" (UID: \"835a5c57-f564-4d91-a2fa-e8b7c8d576b3\") " Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.803652 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-utilities" (OuterVolumeSpecName: "utilities") pod "835a5c57-f564-4d91-a2fa-e8b7c8d576b3" (UID: "835a5c57-f564-4d91-a2fa-e8b7c8d576b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.817576 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-kube-api-access-b74zs" (OuterVolumeSpecName: "kube-api-access-b74zs") pod "835a5c57-f564-4d91-a2fa-e8b7c8d576b3" (UID: "835a5c57-f564-4d91-a2fa-e8b7c8d576b3"). InnerVolumeSpecName "kube-api-access-b74zs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.904166 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "835a5c57-f564-4d91-a2fa-e8b7c8d576b3" (UID: "835a5c57-f564-4d91-a2fa-e8b7c8d576b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.905636 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.905697 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b74zs\" (UniqueName: \"kubernetes.io/projected/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-kube-api-access-b74zs\") on node \"crc\" DevicePath \"\"" Nov 25 10:28:59 crc kubenswrapper[4926]: I1125 10:28:59.905717 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/835a5c57-f564-4d91-a2fa-e8b7c8d576b3-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.545107 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-9r4r6" event={"ID":"835a5c57-f564-4d91-a2fa-e8b7c8d576b3","Type":"ContainerDied","Data":"eee673dbe13d9be2ba6590bbc82bccd78bc624730fdb6621a46b33cd82aeac84"} Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.545182 4926 scope.go:117] "RemoveContainer" containerID="3e5722213257da5c2d6f50d5b0b04c8cb9f2e4068f95d516f9f6846498acba96" Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.545543 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-9r4r6" Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.570724 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-9r4r6"] Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.580467 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-9r4r6"] Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.592392 4926 scope.go:117] "RemoveContainer" containerID="6bd9980837c606738969e00179eac73b25611bb64eca8a67d5193133f5597aa3" Nov 25 10:29:00 crc kubenswrapper[4926]: I1125 10:29:00.773251 4926 scope.go:117] "RemoveContainer" containerID="f5550879b016fd8d2f094d2a8bc8296d8d402bf6eb50c1cd4e86ef4ccd956791" Nov 25 10:29:02 crc kubenswrapper[4926]: I1125 10:29:02.022714 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" path="/var/lib/kubelet/pods/835a5c57-f564-4d91-a2fa-e8b7c8d576b3/volumes" Nov 25 10:29:03 crc kubenswrapper[4926]: I1125 10:29:03.012138 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:29:03 crc kubenswrapper[4926]: E1125 10:29:03.012375 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:29:15 crc kubenswrapper[4926]: I1125 10:29:15.012227 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:29:15 crc kubenswrapper[4926]: E1125 10:29:15.013178 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.399333 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9p6lt"] Nov 25 10:29:21 crc kubenswrapper[4926]: E1125 10:29:21.400151 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="extract-utilities" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.400164 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="extract-utilities" Nov 25 10:29:21 crc kubenswrapper[4926]: E1125 10:29:21.400190 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="extract-content" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.400197 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="extract-content" Nov 25 10:29:21 crc kubenswrapper[4926]: E1125 10:29:21.400230 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="registry-server" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.400235 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="registry-server" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.400407 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="835a5c57-f564-4d91-a2fa-e8b7c8d576b3" containerName="registry-server" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.402026 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.418427 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9p6lt"] Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.578687 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk5vb\" (UniqueName: \"kubernetes.io/projected/eee16b33-ba11-48ea-b2eb-13e30f8ef762-kube-api-access-bk5vb\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.578752 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-utilities\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.578816 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-catalog-content\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.681135 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-utilities\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.681184 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-catalog-content\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.681310 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk5vb\" (UniqueName: \"kubernetes.io/projected/eee16b33-ba11-48ea-b2eb-13e30f8ef762-kube-api-access-bk5vb\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.681737 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-utilities\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.681929 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-catalog-content\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.718177 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk5vb\" (UniqueName: \"kubernetes.io/projected/eee16b33-ba11-48ea-b2eb-13e30f8ef762-kube-api-access-bk5vb\") pod \"certified-operators-9p6lt\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:21 crc kubenswrapper[4926]: I1125 10:29:21.732740 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:22 crc kubenswrapper[4926]: I1125 10:29:22.382383 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9p6lt"] Nov 25 10:29:22 crc kubenswrapper[4926]: I1125 10:29:22.748083 4926 generic.go:334] "Generic (PLEG): container finished" podID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerID="ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db" exitCode=0 Nov 25 10:29:22 crc kubenswrapper[4926]: I1125 10:29:22.748138 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerDied","Data":"ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db"} Nov 25 10:29:22 crc kubenswrapper[4926]: I1125 10:29:22.748179 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerStarted","Data":"f6722f500f6b698d5f852e66503123811c9d5da20ff43d75e5b8ed208e0965ec"} Nov 25 10:29:24 crc kubenswrapper[4926]: I1125 10:29:24.770614 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerStarted","Data":"444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c"} Nov 25 10:29:25 crc kubenswrapper[4926]: I1125 10:29:25.781638 4926 generic.go:334] "Generic (PLEG): container finished" podID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerID="444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c" exitCode=0 Nov 25 10:29:25 crc kubenswrapper[4926]: I1125 10:29:25.781678 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerDied","Data":"444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c"} Nov 25 10:29:27 crc kubenswrapper[4926]: I1125 10:29:27.012887 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:29:27 crc kubenswrapper[4926]: E1125 10:29:27.013435 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:29:27 crc kubenswrapper[4926]: I1125 10:29:27.811865 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerStarted","Data":"aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d"} Nov 25 10:29:27 crc kubenswrapper[4926]: I1125 10:29:27.843363 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9p6lt" podStartSLOduration=2.692960137 podStartE2EDuration="6.843341333s" podCreationTimestamp="2025-11-25 10:29:21 +0000 UTC" firstStartedPulling="2025-11-25 10:29:22.750012878 +0000 UTC m=+2459.433969305" lastFinishedPulling="2025-11-25 10:29:26.900394074 +0000 UTC m=+2463.584350501" observedRunningTime="2025-11-25 10:29:27.827428764 +0000 UTC m=+2464.511385191" watchObservedRunningTime="2025-11-25 10:29:27.843341333 +0000 UTC m=+2464.527297760" Nov 25 10:29:31 crc kubenswrapper[4926]: I1125 10:29:31.733210 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:31 crc kubenswrapper[4926]: I1125 10:29:31.733542 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:31 crc kubenswrapper[4926]: I1125 10:29:31.783355 4926 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:41 crc kubenswrapper[4926]: I1125 10:29:41.778056 4926 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:41 crc kubenswrapper[4926]: I1125 10:29:41.823470 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9p6lt"] Nov 25 10:29:41 crc kubenswrapper[4926]: I1125 10:29:41.939758 4926 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9p6lt" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="registry-server" containerID="cri-o://aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d" gracePeriod=2 Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.012045 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:29:42 crc kubenswrapper[4926]: E1125 10:29:42.012341 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.443095 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.604316 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-catalog-content\") pod \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.604457 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bk5vb\" (UniqueName: \"kubernetes.io/projected/eee16b33-ba11-48ea-b2eb-13e30f8ef762-kube-api-access-bk5vb\") pod \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.604578 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-utilities\") pod \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\" (UID: \"eee16b33-ba11-48ea-b2eb-13e30f8ef762\") " Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.605915 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-utilities" (OuterVolumeSpecName: "utilities") pod "eee16b33-ba11-48ea-b2eb-13e30f8ef762" (UID: "eee16b33-ba11-48ea-b2eb-13e30f8ef762"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.609666 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eee16b33-ba11-48ea-b2eb-13e30f8ef762-kube-api-access-bk5vb" (OuterVolumeSpecName: "kube-api-access-bk5vb") pod "eee16b33-ba11-48ea-b2eb-13e30f8ef762" (UID: "eee16b33-ba11-48ea-b2eb-13e30f8ef762"). InnerVolumeSpecName "kube-api-access-bk5vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.653918 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eee16b33-ba11-48ea-b2eb-13e30f8ef762" (UID: "eee16b33-ba11-48ea-b2eb-13e30f8ef762"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.707399 4926 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-utilities\") on node \"crc\" DevicePath \"\"" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.707432 4926 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eee16b33-ba11-48ea-b2eb-13e30f8ef762-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.707446 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bk5vb\" (UniqueName: \"kubernetes.io/projected/eee16b33-ba11-48ea-b2eb-13e30f8ef762-kube-api-access-bk5vb\") on node \"crc\" DevicePath \"\"" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.954440 4926 generic.go:334] "Generic (PLEG): container finished" podID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerID="aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d" exitCode=0 Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.954507 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerDied","Data":"aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d"} Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.954577 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9p6lt" event={"ID":"eee16b33-ba11-48ea-b2eb-13e30f8ef762","Type":"ContainerDied","Data":"f6722f500f6b698d5f852e66503123811c9d5da20ff43d75e5b8ed208e0965ec"} Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.954604 4926 scope.go:117] "RemoveContainer" containerID="aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.954521 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9p6lt" Nov 25 10:29:42 crc kubenswrapper[4926]: I1125 10:29:42.980012 4926 scope.go:117] "RemoveContainer" containerID="444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.003222 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9p6lt"] Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.004868 4926 scope.go:117] "RemoveContainer" containerID="ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.013212 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9p6lt"] Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.048468 4926 scope.go:117] "RemoveContainer" containerID="aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d" Nov 25 10:29:43 crc kubenswrapper[4926]: E1125 10:29:43.048983 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d\": container with ID starting with aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d not found: ID does not exist" containerID="aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.049015 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d"} err="failed to get container status \"aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d\": rpc error: code = NotFound desc = could not find container \"aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d\": container with ID starting with aad098862fc13777c71b75983fb7e19d2821a5e68e7f5806e43e0ee9a18e6f9d not found: ID does not exist" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.049048 4926 scope.go:117] "RemoveContainer" containerID="444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c" Nov 25 10:29:43 crc kubenswrapper[4926]: E1125 10:29:43.049547 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c\": container with ID starting with 444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c not found: ID does not exist" containerID="444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.049595 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c"} err="failed to get container status \"444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c\": rpc error: code = NotFound desc = could not find container \"444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c\": container with ID starting with 444fcf7b6ae61283288f3816947cc5ab8952bf9bee2f5ec3d3b9a551eb5c213c not found: ID does not exist" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.049628 4926 scope.go:117] "RemoveContainer" containerID="ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db" Nov 25 10:29:43 crc kubenswrapper[4926]: E1125 10:29:43.050008 4926 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db\": container with ID starting with ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db not found: ID does not exist" containerID="ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db" Nov 25 10:29:43 crc kubenswrapper[4926]: I1125 10:29:43.050032 4926 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db"} err="failed to get container status \"ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db\": rpc error: code = NotFound desc = could not find container \"ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db\": container with ID starting with ced96d09f032fb70020fbbb580beb55c71b6259983b470460e311eaab31135db not found: ID does not exist" Nov 25 10:29:44 crc kubenswrapper[4926]: I1125 10:29:44.022257 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" path="/var/lib/kubelet/pods/eee16b33-ba11-48ea-b2eb-13e30f8ef762/volumes" Nov 25 10:29:54 crc kubenswrapper[4926]: I1125 10:29:54.018188 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:29:54 crc kubenswrapper[4926]: E1125 10:29:54.018985 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.188049 4926 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd"] Nov 25 10:30:00 crc kubenswrapper[4926]: E1125 10:30:00.189199 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="extract-content" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.189218 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="extract-content" Nov 25 10:30:00 crc kubenswrapper[4926]: E1125 10:30:00.189236 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="registry-server" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.189242 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="registry-server" Nov 25 10:30:00 crc kubenswrapper[4926]: E1125 10:30:00.189271 4926 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="extract-utilities" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.189278 4926 state_mem.go:107] "Deleted CPUSet assignment" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="extract-utilities" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.189467 4926 memory_manager.go:354] "RemoveStaleState removing state" podUID="eee16b33-ba11-48ea-b2eb-13e30f8ef762" containerName="registry-server" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.190113 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.192926 4926 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.193441 4926 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.196997 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd"] Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.231922 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c438cc-edab-40d1-8c06-6b0b1915cc78-config-volume\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.232353 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cb44\" (UniqueName: \"kubernetes.io/projected/a3c438cc-edab-40d1-8c06-6b0b1915cc78-kube-api-access-4cb44\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.232628 4926 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c438cc-edab-40d1-8c06-6b0b1915cc78-secret-volume\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.333870 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cb44\" (UniqueName: \"kubernetes.io/projected/a3c438cc-edab-40d1-8c06-6b0b1915cc78-kube-api-access-4cb44\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.333993 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c438cc-edab-40d1-8c06-6b0b1915cc78-secret-volume\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.334052 4926 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c438cc-edab-40d1-8c06-6b0b1915cc78-config-volume\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.335252 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c438cc-edab-40d1-8c06-6b0b1915cc78-config-volume\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.346166 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c438cc-edab-40d1-8c06-6b0b1915cc78-secret-volume\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.353414 4926 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cb44\" (UniqueName: \"kubernetes.io/projected/a3c438cc-edab-40d1-8c06-6b0b1915cc78-kube-api-access-4cb44\") pod \"collect-profiles-29401110-nh8gd\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.518990 4926 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:00 crc kubenswrapper[4926]: I1125 10:30:00.969284 4926 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd"] Nov 25 10:30:01 crc kubenswrapper[4926]: I1125 10:30:01.120125 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" event={"ID":"a3c438cc-edab-40d1-8c06-6b0b1915cc78","Type":"ContainerStarted","Data":"631591cc6e95fa0bd99a15ea140784b7ed08748add0e4d26467e25e67e6cc7c6"} Nov 25 10:30:01 crc kubenswrapper[4926]: I1125 10:30:01.120545 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" event={"ID":"a3c438cc-edab-40d1-8c06-6b0b1915cc78","Type":"ContainerStarted","Data":"cff3afae5d90fc32300c25651189d919c838733c6abbc5fef934a20599cf8174"} Nov 25 10:30:01 crc kubenswrapper[4926]: I1125 10:30:01.141650 4926 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" podStartSLOduration=1.141633695 podStartE2EDuration="1.141633695s" podCreationTimestamp="2025-11-25 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-25 10:30:01.134331772 +0000 UTC m=+2497.818288219" watchObservedRunningTime="2025-11-25 10:30:01.141633695 +0000 UTC m=+2497.825590112" Nov 25 10:30:02 crc kubenswrapper[4926]: I1125 10:30:02.128520 4926 generic.go:334] "Generic (PLEG): container finished" podID="a3c438cc-edab-40d1-8c06-6b0b1915cc78" containerID="631591cc6e95fa0bd99a15ea140784b7ed08748add0e4d26467e25e67e6cc7c6" exitCode=0 Nov 25 10:30:02 crc kubenswrapper[4926]: I1125 10:30:02.128576 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" event={"ID":"a3c438cc-edab-40d1-8c06-6b0b1915cc78","Type":"ContainerDied","Data":"631591cc6e95fa0bd99a15ea140784b7ed08748add0e4d26467e25e67e6cc7c6"} Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:03.455247 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:03.488511 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cb44\" (UniqueName: \"kubernetes.io/projected/a3c438cc-edab-40d1-8c06-6b0b1915cc78-kube-api-access-4cb44\") pod \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:03.488610 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c438cc-edab-40d1-8c06-6b0b1915cc78-config-volume\") pod \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:03.488722 4926 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c438cc-edab-40d1-8c06-6b0b1915cc78-secret-volume\") pod \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\" (UID: \"a3c438cc-edab-40d1-8c06-6b0b1915cc78\") " Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:03.489953 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3c438cc-edab-40d1-8c06-6b0b1915cc78-config-volume" (OuterVolumeSpecName: "config-volume") pod "a3c438cc-edab-40d1-8c06-6b0b1915cc78" (UID: "a3c438cc-edab-40d1-8c06-6b0b1915cc78"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:03.795025 4926 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a3c438cc-edab-40d1-8c06-6b0b1915cc78-config-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.131392 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3c438cc-edab-40d1-8c06-6b0b1915cc78-kube-api-access-4cb44" (OuterVolumeSpecName: "kube-api-access-4cb44") pod "a3c438cc-edab-40d1-8c06-6b0b1915cc78" (UID: "a3c438cc-edab-40d1-8c06-6b0b1915cc78"). InnerVolumeSpecName "kube-api-access-4cb44". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.137745 4926 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a3c438cc-edab-40d1-8c06-6b0b1915cc78-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a3c438cc-edab-40d1-8c06-6b0b1915cc78" (UID: "a3c438cc-edab-40d1-8c06-6b0b1915cc78"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.183314 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" event={"ID":"a3c438cc-edab-40d1-8c06-6b0b1915cc78","Type":"ContainerDied","Data":"cff3afae5d90fc32300c25651189d919c838733c6abbc5fef934a20599cf8174"} Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.183369 4926 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cff3afae5d90fc32300c25651189d919c838733c6abbc5fef934a20599cf8174" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.183463 4926 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29401110-nh8gd" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.229353 4926 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cb44\" (UniqueName: \"kubernetes.io/projected/a3c438cc-edab-40d1-8c06-6b0b1915cc78-kube-api-access-4cb44\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.229383 4926 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a3c438cc-edab-40d1-8c06-6b0b1915cc78-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.532291 4926 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq"] Nov 25 10:30:04 crc kubenswrapper[4926]: I1125 10:30:04.541070 4926 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29401065-qbgkq"] Nov 25 10:30:06 crc kubenswrapper[4926]: I1125 10:30:06.025038 4926 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd0f7444-df82-40cd-aaa3-60250afa1f36" path="/var/lib/kubelet/pods/dd0f7444-df82-40cd-aaa3-60250afa1f36/volumes" Nov 25 10:30:08 crc kubenswrapper[4926]: I1125 10:30:08.014689 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:30:08 crc kubenswrapper[4926]: E1125 10:30:08.015614 4926 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-b82dg_openshift-machine-config-operator(7142d4cf-9f77-4d00-be33-0198a4cb84d1)\"" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" podUID="7142d4cf-9f77-4d00-be33-0198a4cb84d1" Nov 25 10:30:21 crc kubenswrapper[4926]: I1125 10:30:21.012524 4926 scope.go:117] "RemoveContainer" containerID="eb6f17cc61954e8e7f4bbea99773ac2ec4c3edf342983573f4eb6093dd3fca28" Nov 25 10:30:21 crc kubenswrapper[4926]: I1125 10:30:21.318263 4926 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-b82dg" event={"ID":"7142d4cf-9f77-4d00-be33-0198a4cb84d1","Type":"ContainerStarted","Data":"764d176eaa7ed29f59adc3f1297daf29846540f5c415315d16e0c061053fe5f9"} Nov 25 10:30:38 crc kubenswrapper[4926]: I1125 10:30:38.097060 4926 scope.go:117] "RemoveContainer" containerID="5aa5da072b249b90372df24f403b9372abe3d2aead066774905eea565d65b663" Nov 25 10:30:38 crc kubenswrapper[4926]: I1125 10:30:38.119421 4926 scope.go:117] "RemoveContainer" containerID="9ea075134112737ea4398b1fd34e52d0e48fcec09fee71009604c69d9813f247" Nov 25 10:30:38 crc kubenswrapper[4926]: I1125 10:30:38.174472 4926 scope.go:117] "RemoveContainer" containerID="fec6178f6cd2488782d41d0ddafb2f6ab1df8915b1a6d78c6257c16ffc187a71" Nov 25 10:30:38 crc kubenswrapper[4926]: I1125 10:30:38.215452 4926 scope.go:117] "RemoveContainer" containerID="8c6543c4e60c937ae9f30ace8193728c6107dd73a19f107096f7dd06a4d274ef" Nov 25 10:30:38 crc kubenswrapper[4926]: I1125 10:30:38.239502 4926 scope.go:117] "RemoveContainer" containerID="42a8de03ebbfdbb0aeaaa9cf0d2568fc81055cae01217b924427459ee714959e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111302556024444 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111302556017361 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111275235016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111275235015457 5ustar corecore